Description
See Gecko Bug 1661368 and more recently Bug 1814791.
There is currently no validation of whether a file with an stsc
box with a large sample count is actually referencing offsets that fall within the file, which allows create_sample_table()
to allocate an arbitrarily large vector and OOM.
We could just set a configurable hard limit on the number of samples it can allocate, but I don't know if that's the best solution.
We can also fail early when the relevant sample boxes have mismatched sample counts, but I'm not sure if there are files that currently parse and break that rule, and as tnikkel pointed out, the fuzzer may still find a way to create a matching sample count that causes OOM. Might still be worth implementing this if it makes sense to fail that way.
I think the most ideal solution is instead to make create_sample_table()
aware of the amount of data in the file that is currently available to the caller, and use the stsz
box to determine how many samples the vector needs to allocate to reach EOF. For cases where the full size is not known beforehand, API users would have to call get_indices
multiple times and be aware that their pointers will be invalid, but that doesn't seem like a difficult problem.
What I wonder is whether the available data is something that should be accessible from other parts of mp4parse as well, for example to allow the box parsing to indicate that it hit the end of the available data but can continue parsing when more data is available.
Feel free to let me know if this idea doesn't make sense, we can use this issue to brainstorm more solutions if necessary.