10000 Batch size affects model's output · Issue #249 · ggml-org/llama.cpp · GitHub
[go: up one dir, main page]

Skip to content
Batch size affects model's output #249
Closed
@oKatanaaa

Description

@oKatanaaa

I was tinkering with the code and made the following change in line 977, main.cpp (as it seemed wrong to me):
from

if (embd.size() > params.n_batch) {
       break;
}

to

if (embd.size() >= params.n_batch) {
       break;
}

The model's (13B) outputs suddenly changed. Reverted changes and tried to play with the batch_size parameter, it really does affect the output.

Not sure if it's expected behaviour. As far as I understand it shouldn't be the case. A bug? Different batch sizes have different evaluation results (rounding error)?

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't workinggeneration qualityQuality of model output

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Issue actions

      0