[tools/data_processing_v2] - Make Processing Data More User Friendly · Issue #368 · waterloo-rocketry/omnibus · GitHub
More Web Proxy on the site http://driver.im/
You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Expected Behavior: I would want to be able to have a nice interface / CLI to use to process globallog log files into CSVs. Furthermore, when processing the msgpacked log files, the entire log should be processed line by line and shouldn't be read into memory before processing.
Because: A proper CLI / UI would make data processing a lot more user friendly while retaining the new features. Making the code not reliant on memory space reduces memory requirements and prevents most OOM issues. Therefore, it would allow for very large logs to processed without fuss (and then processed in chunks later in Pandas).
The script MUST provide a basic CLI that allows the user to run it as python main.py daq [source file] and output to a CSV file in the same folder and with a default file name + a random hash at the end (example: omnibus-processed-daq-[date]-[hash].csv).
All DAQ functions MUST be in a daq subcommand (as seen above) to distinguish from future log parsing functionality.
The script SHOULD provide a CLI option -o [outfile] to specify the output file name
The script MUST process the msgpacked log line by line, do not read into a single list
Standard Requirements - Do Not Modify Without Approval
All changes made MUST be tested and reviewed by a code reviewer.
The conditions outlined in CONTRIBUTING.MD MUST be respected.
Deep consider of new library dependencies MUST be made before adding them to the project.
Keep the codebase clean, don't push non-code files (e.g. DATA, VirtualEnv, etc.) to the repository.
All new feature changes made SHOULD have associated unit tests (if applicable)
Keep well-documented code, and follow the style guide.
Satisfying the above requirements WILL close this issue.
ChrisYx511
changed the title
[tools/data_processing_v2] - New Data Processing Script!
[tools/data_processing_v2] - Make Processing Data More User Friendly
Mar 18, 2025
Uh oh!
There was an error while loading. Please reload this page.
Feature Description
Expected implementation date: 2025-03-29
As a: DAQ Ops,
Expected Behavior: I would want to be able to have a nice interface / CLI to use to process globallog log files into CSVs. Furthermore, when processing the msgpacked log files, the entire log should be processed line by line and shouldn't be read into memory before processing.
Because: A proper CLI / UI would make data processing a lot more user friendly while retaining the new features. Making the code not reliant on memory space reduces memory requirements and prevents most OOM issues. Therefore, it would allow for very large logs to processed without fuss (and then processed in chunks later in Pandas).
Links to related issues / PRs / documents / discussions (Slack): https://docs.google.com/document/d/1G8s8M530FyW4kbQcjkgZ2ImOGnDQj_MT3Omr9mVWqEw/edit?tab=t.0#heading=h.watkmfcldmtw
Related: #212
Requirements
Specific Requirements
python main.py daq [source file]
and output to a CSV file in the same folder and with a default file name + a random hash at the end (example:omnibus-processed-daq-[date]-[hash].csv
).daq
subcommand (as seen above) to distinguish from future log parsing functionality.-o [outfile]
to specify the output file nameStandard Requirements - Do Not Modify Without Approval
Satisfying the above requirements WILL close this issue.
Stakeholder / Product Owner
DAQ Ops
Mention: @ChrisYx511
The text was updated successfully, but these errors were encountered: