Guoqing Zheng Profile
Guoqing Zheng

@zzzzgq

Followers
106
Following
87
Media
1
Statuses
33

Research Scientist at Percepta AI | Ex: Principal Researcher at @MSFTResearch.

Joined March 2009
Don't wanna be here? Send us removal request.
@zzzzgq
Guoqing Zheng
11 days
Join us at NeurIPS for a happy hour on AI transformation, co-hosted by Percepta, GC and @nvidia: RSVP at https://t.co/6af38RxjlJ, and solve some fun puzzles at https://t.co/kaKAK2ECYu to skip the waiting list!
Tweet card summary image
invite.generalcatalyst.com
Hosted by Percepta, General Catalyst, NVIDIA
0
2
6
@apjacob03
Athul Paul Jacob
2 months
Today marks an important milestone. I’m launching Percepta together with @htaneja, @hirshjain, @tmathew0309, Radha Jain, @marisbest2, @KonstDaskalakis and an incredible team, with the goal of bringing AI to the core industries that run our economy. For AI to deliver
percepta.ai
Transforming critical institutions using applied AI. Let's harness the frontier.
10
20
86
@mojan_jp
Mojan Javaheripi
7 months
Excited to release our first set of reasoning models Phi-4-reasoning and Phi-4-reasoning-plus, available today on HuggingFace and Azure AI foundry. Some interesting insights below and more deep dives in following days!
1
10
42
@AhmedHAwadallah
Ahmed Awadallah
7 months
Introducing Phi-4-reasoning, adding reasoning models to the Phi family of SLMs. The model is trained with both supervised finetuning (using a carefully curated dataset of reasoning demonstration) and Reinforcement Learning. 📌Competitive results on reasoning benchmarks with
4
34
141
@DimitrisPapail
Dimitris Papailiopoulos
7 months
We’ve been cooking... a new open weights 14B Phi-4 reasoning model, SFT’d on ~1.4M carefully curated reasoning demonstrations from o3-mini and RL’d for a tiny bit. This model is a little beast.
36
239
1K
@zzzzgq
Guoqing Zheng
1 year
We are hiring interns to work on LLM reasoning for summer 2025! Apply at
@DimitrisPapail
Dimitris Papailiopoulos
1 year
How do you train reasoning models? What's the role of verifiers, RL, and synth data generation? How do these fit in multi-agent workflows? To find out, come join us for an internship at MSR AI Frontiers. Link below :D
0
1
8
@DimitrisPapail
Dimitris Papailiopoulos
1 year
What is reasoning? Do LLMs use it? Does it help? Is o1 really that better than sonnet? How do you even measure all that? MSR AI Frontiers is working to figure it all out, and we're looking for interns to work on evals to better understand LLMs. Please apply!! Link below:
4
8
143
@Arindam1408
arindam mitra
1 year
#Orca I'm thrilled to announce our latest work on Generative Teaching: generating vast amount of diverse high-quality synthetic data for language models to teach a specific skill (e.g. RC, text classification, tool use,math) without the extensive human effort typically required
4
9
44
@Arindam1408
arindam mitra
2 years
With Orca, we're excited about the potential of redefining the reasoning capabilities of smaller LLMs. We're still at the beginning phases of this intriguing journey, but our preliminary explorations have yielded encouraging results. 1/7
5
13
40
@zzzzgq
Guoqing Zheng
4 years
Cleanly labeled and weakly labeled data are both crucial to few-shot NLU. WALNUT features a unified setting with both few-shot and weakly supervised learning. We hope to bring more attention to semi-weakly supervised learning for NLU. #NAACL2022 #NLProc @MSFTResearch
@KaiShu0327
Kai Shu
4 years
Our paper entitled "WALNUT: A Benchmark on Semi-weakly Supervised Learning for Natural Language Understanding" got accepted in #NAACL2022. Joint with @zzzzgq @gkaraml and @AhmedHAwadallah, from Microsoft Research and Columbia. Stay tuned!
0
0
2
@zzzzgq
Guoqing Zheng
4 years
paper: https://t.co/1Gp0GIbtlF, code: https://t.co/oCsNM1WLMP. Joint work fom Microsoft, Dartmouth and Northwestern.
Tweet card summary image
github.com
Knowledge Infused Decoding. Contribute to microsoft/KID development by creating an account on GitHub.
@zzzzgq
Guoqing Zheng
4 years
Integrating knowledge to NLG typically requires training/fine-tuning PLMs on knowledge sources. With KID, we show it's viable to infuse knowledge on the fly in decoding phase without requiring modifications to the LMs. Check out our #iclr2022 work on Knowledge Infused Decoding.
1
1
1
@zzzzgq
Guoqing Zheng
4 years
Integrating knowledge to NLG typically requires training/fine-tuning PLMs on knowledge sources. With KID, we show it's viable to infuse knowledge on the fly in decoding phase without requiring modifications to the LMs. Check out our #iclr2022 work on Knowledge Infused Decoding.
0
1
4
@zzzzgq
Guoqing Zheng
4 years
Two new NLU benchmarks for robust evaluation of PLMs.
@MSFTResearch
Microsoft Research
4 years
Current benchmarks may yield imprecise readings of AI models’ natural language understanding. Two new NLU benchmarks aim for more accurate evaluations. #NeurIPS2021
0
0
1
@subho_mpi
Subhabrata Mukherjee
4 years
Open to everyone! The first-ever 2021 Microsoft Research Summit, Oct 19 - 21, with over 150 sessions across 16 tracks, provides the global research community with an opportunity learn from experts pushing the frontiers of technology. Register now:… https://t.co/cK5qBSomc4
Tweet card summary image
linkedin.com
Open to everyone! The first-ever 2021 Microsoft Research Summit, Oct 19 - 21, with over 150 sessions across 16 tracks, provides the global research community with an opportunity learn from experts...
0
3
7
@gneubig
Graham Neubig
4 years
Just released a new survey on prompting methods, which use language models to solve prediction tasks by providing them with a "prompt" like: "CMU is located in __" We worked really hard to make this well-organized and educational for both NLP experts and beginners, check it out!
@stefan_fee
Pengfei Liu
4 years
What is prompt-based learning, and what challenges are there? Will it be a new paradigm or a way for human-PLMs communication? How does it connect with other research and how to position it in the evolution of the NLP research paradigm? We released a systematic survey and beyond
1
54
216
@zzzzgq
Guoqing Zheng
5 years
Excited to share our recent work on leveraging meta-learning for cross-lingual transfer.
@gneubig
Graham Neubig
5 years
MetaXL is a new method for cross-lingual transfer to extremely low-resource languages that works by meta-learning transformation functions to improve gradient alignment between source and target languages. See our #NAACL2021 paper! https://t.co/tf61zMEjQ8 1/2
0
0
2
@zzzzgq
Guoqing Zheng
12 years
W
0
0
1
@nikete
nikete
12 years
An Equivalence between the Lasso and Support Vector Machines http://t.co/31go7AqTTl
0
6
14