You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Following a successful research bounty into the OneDiff optimization algorithm, we've discovered that combining OneDiff with OneFlow can accelerate inference by approximately 60% while maintaining image quality. The only drawback is an increase in startup time from 2 seconds to 68 seconds. Given that a 60% reduction in inference time can significantly enhance network performance, we believe it's crucial to provide orchestrators with the means to enable this optimization.
Currently, the core AI team is occupied with major restructuring tasks, including the AI worker and pipeline generalisations, developing new pipelines for startups, and releasing client SDKs. Therefore, we are contacting the community to implement this vital optimization for us 🏗️⚡.
Implement a mechanism for orchestrators to enable the OneDiff optimization with the OneFlow compiler on the T2I pipeline using a ONEDIFF Environment variable, similar to the approach used for Deep Cache and Stable Fast optimizations. For reference, see the implementation here.
Integration in AI-Runner Container:
Ensure that the OneDiff optimization can be integrated into the ai-runner container without causing dependency conflicts with existing pipelines or optimizations. This will likely involve updating Torch.
Implementation Tips
Utilize Developer Documentation:
Refer to our developer documentation for both the worker and runner. These resources offer valuable tips for accelerating your development process, including how to mock pipelines and enable direct debugging.
Review Existing Optimizations:
The T2I pipeline already includes several optimizations. Review the existing code to understand these optimizations and incorporate similar strategies when implementing the new OneDiff optimization.
Seek Help Promptly:
This bounty is challenging and may involve dependency errors that the core AI team has likely encountered before. If you get stuck, don’t hesitate to ask questions in the #🛋│developer-lounge channel.
How to Apply
Express Your Interest: Comment on this issue to indicate your interest and explain why you're the ideal candidate for the task.
Wait for Review: Our team will review expressions of interest and select the best candidate.
Get Assigned: If selected, we'll assign the GitHub issue to you.
Start Working: Dive into your task! If you need assistance or guidance, comment on the issue or join the discussions in the #🛋│developer-lounge channel on our Discord server.
Submit Your Work: Create a pull request in the relevant repository and request a review.
Notify Us: Comment on this GitHub issue when your pull request is ready for review.
Receive Your Bounty: We'll arrange the bounty payment once your pull request is approved.
Gain Recognition: Your valuable contributions will be showcased in our project's changelog.
Thank you for your interest in contributing to our project 💛!
Warning
Please wait for the issue to be assigned to you before starting work. To prevent duplication of effort, submissions for unassigned issues will not be accepted.
The text was updated successfully, but these errors were encountered:
@lukiod, it's great to see that you've decided to take on the implementation following the resolution of #31 🚀. Given your experience with the OneDiff optimization in your research, I believe you're the ideal candidate for this task. I've assigned this bounty to you accordingly.
Overview
Following a successful research bounty into the OneDiff optimization algorithm, we've discovered that combining OneDiff with OneFlow can accelerate inference by approximately 60% while maintaining image quality. The only drawback is an increase in startup time from 2 seconds to 68 seconds. Given that a 60% reduction in inference time can significantly enhance network performance, we believe it's crucial to provide orchestrators with the means to enable this optimization.
Currently, the core AI team is occupied with major restructuring tasks, including the AI worker and pipeline generalisations, developing new pipelines for startups, and releasing client SDKs. Therefore, we are contacting the community to implement this vital optimization for us 🏗️⚡.
Required Skillset
Bounty Requirements
ONEDIFF
Environment variable, similar to the approach used for Deep Cache and Stable Fast optimizations. For reference, see the implementation here.Implementation Tips
How to Apply
Thank you for your interest in contributing to our project 💛!
Warning
Please wait for the issue to be assigned to you before starting work. To prevent duplication of effort, submissions for unassigned issues will not be accepted.
The text was updated successfully, but these errors were encountered: