Quantum_Stat Profile Banner
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝ Profile
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝

@Quantum_Stat

Followers
2K
Following
542
Media
334
Statuses
4K

҈̿ ̶̢̢̧̡̼̜̝̬͍̜̘̙͉̘͎͓͍̣̰͖̹͖͚̭̘̖̟͕̬̠̬͇̹̮͎̣̱̎̾͌̊́̇͛̅͂̀̀͑̃̈̀̓̏͌͌͋̐̾̒͋͋̏͋̽̈́͐͑̐̂͆̊̈́̾͌̓͌̕̚͝͝͠͠͝ͅ Repos: https://t.co/3x0DctKuld

Joined March 2018
Don't wanna be here? Send us removal request.
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
🚀🚀 Super excited to share the latest benchmark results for our quantized BGE models. A few weeks ago, these models were introduced with the aim of enhancing performance and efficiency for generating embeddings. And we've now conducted thorough comparisons between running
0
2
10
@rajkbnp
Raj KB
2 years
I love the #ChatGPT Cheat Sheet by Ricky Costa (@Quantum_Stat) which includes 🔹NLP Tasks 🔹Code 🔹Structured Output Styles 🔹Unstructured Output Styles 🔹Media Types 🔹Meta ChatGPT 🔹Expert Prompting Get your hands on this amazing resource at: https://t.co/Bg1roxcMFO
0
1
4
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
0
0
0
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
GUIDE:
0
0
0
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
Check the image below for an example of what I'm discussing 👇 We are soon releasing a notebook with an end-to-end example for anyone to replicate the compressed bge models which achieve great accuracy results on the MTEB Leaderboard.
1
0
0
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
The .npz file is a dictionary, with keys mapping to input names in the ONNX spec and values as NumPy arrays filled with the data. - Keep all data samples in a directory, usually named "data."
1
0
0
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
📦 The NPZ files will be used by Sparsify to calibrate the ONNX model by using samples from a calibration dataset. 🔍 **Specifications**: - Each .npz file houses a single data sample, no batch dimension. This data sample takes a thrilling ride through the ONNX model. -
1
0
0
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
⚡Getting to Know the NPZ file format to Compress BGE Embedding Models ⚡ For One-Shot Quantization (INT8), Sparsify relies on the .npz format for data storage, a file format rooted in the mighty NumPy library.
2
1
4
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
source library Sparsify! Not only is it ONNX and INT8 quantized (faster and lighter) but is able to run on CPUs using DeepSparse! 💥 cc @neuralmagic Model:
Tweet card summary image
huggingface.co
0
0
7
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
⚡IT HAPPENED!⚡ There's a new state-of-the-art sentence embeddings model for the semantic textual similarity task on Hugging Face's MTEB leaderboard 🤗! Bge-large-en-v1.5-quant was the model I quantized in less than an hour using a single CLI command using Neural Magic's open
3
6
45
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
Exciting News! 🚀 DeepSparse is now integrated with @LangChainAI , opening up a world of possibilities in Generative AI on CPUs. Langchain, known for its innovative design paradigms for large language model (LLM) applications, was often constrained by expensive APIs or cumbersome
0
6
22
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
🌟First, want to thank everyone for pushing this model past 1,000 downloads in only a few days!! Additionally, I added bge-base models to MTEB. Most importantly, code snippets were added for running inference in the model cards for everyone to try out! https://t.co/NZO7DPGubb
Tweet card summary image
huggingface.co
0
0
5
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
lol
@greg16676935420
greg
2 years
Numbers 1-100 ranked, worst to best 100. 39 99. 41 98. 8 97. 43 96. 59 95. 74 94. 61 93. 89 92. 58 91. 12 90. 14 89. 19 88. 38 87. 71 86. 73 85. 55 84. 56 83. 37 82. 45 81. 76 80. 78 79. 96 78. 98 77. 87 76. 6 75. 68 74. 79 73. 85 72. 34 71. 42 70. 25 69. 18 68. 5 67. 84 66. 83
0
0
1
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
🚀🚀 Explore Sparsify's One-Shot Experiment Guide! Discover how to quickly optimize your models with post-training algorithms for a 3-5x speedup. Perfect for when you need to sparsify your model with limited time and improved inference speedups.🔥 **FYI, this is what I used to
0
1
5
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
🚀🚀 Hey, check out our blog on @huggingface 🤗regarding running LLMs on CPUs! The blog discusses how researchers at IST Austria & Neural Magic have cracked the code for fine-tuning large language models. The method, combining sparse fine-tuning and distillation-type losses,
0
1
6
@Quantum_Stat
̷̨̨̛̛̲͎̣̞̘͈̯͚͂̈́̂̄̽̎́̽̔͑̄̏̽̏͒̾́̅̐̈́̾̎̆͆̽́͌̽̀̚̕̚̕͠͝͝
2 years
🚀✨ Run CodeGen on CPUs with this detailed Colab notebook! 📝 Explore how to sparsify and perform Large Language Model (LLM) inference using Neural Magic's stack, featuring Salesforce/codegen-350M-mono as an example. Dive into these key steps: 1️⃣ **Installation**: Quickly set
0
1
3