Welcome to ecocompute-dynamic-eval Discussions! #4
Pinned
hongping-zh
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
👋 Welcome to EcoCompute Discussions!
Thank you for your interest in energy-efficient LLM inference research!
🎯 What is this project?
We discovered that default bitsandbytes INT8 uses 17-147% MORE energy than FP16, contrary to common belief. Through systematic ablation experiments across three NVIDIA GPU architectures, we identified mixed-precision decomposition as the root cause.
🔥 Key Findings
Root cause: INT8↔FP16 type conversions in outlier-aware decomposition, not INT8 arithmetic itself.
📊 Research Quality
📁 View Metadata →
📊 Interactive Dashboard →
🤝 How to Participate
💡 Share Ideas
Have suggestions for new experiments or visualizations? Start a discussion in Ideas!
🙋 Ask Questions
Confused about the methodology or results? Ask in Q&A and I'll respond within 24-48 hours.
📊 Share Your Results
Run benchmarks on your GPU and share findings in Results Sharing. We especially need:
🎓 Academic Discussion
Discuss methodology, statistical approaches, or related research in Research.
🤝 Find Collaborators
Looking for co-authors for extended studies? Post in Collaboration!
🚀 Quick Links
📧 Contact
Author: Hongping Zhang
Email: zhanghongping1982@gmail.com
Location: Changsha, Hunan, China
🌟 Current Focus
I'm actively working on:
💬 Discussion Guidelines
Looking forward to your contributions and discussions! 🚀
— Hongping Zhang
"Measure, don't assume. Reproduce, don't trust. Share, don't hoard."
Beta Was this translation helpful? Give feedback.
All reactions