prb977 Profile Banner
Prabin Bhandari Profile
Prabin Bhandari

@prb977

Followers
502
Following
8K
Media
92
Statuses
3K

I do NLP related stuffs.

Fairfax, VA
Joined April 2011
Don't wanna be here? Send us removal request.
@prb977
Prabin Bhandari
1 year
Travel survey data are vital for urban mobility assessments—but they’re often costly and difficult to collect. Could LLMs help us synthesize such data? Our latest work, in collaboration with @anas_ant and @dpfoser, and the Best Paper Winner🥇at #SIGSPATIAL24, shows they can! 🧵👇
2
3
13
@prb977
Prabin Bhandari
1 year
Work done in collaboration with @Harrod_Karlyn and @anas_ant . Read more in our paper: https://t.co/ketmTaiULJ 📄
0
0
0
@prb977
Prabin Bhandari
1 year
Our geotagging results show that LLMs with access to global region databases enable precise geotagging to finer administrative levels.
1
0
0
@prb977
Prabin Bhandari
1 year
Our human evaluations reveal that while 55% of the extracted data contain errors, the LLMs still produce over three times more accurate data than manual extraction methods.
1
0
0
@prb977
Prabin Bhandari
1 year
We tested our approach on Rift Valley Fever data across 22+ countries, evaluating accuracy in extracting outbreak details and assigning correct geolocations. Our automatic evals show that LLMs extract more data than humans.
1
0
0
@prb977
Prabin Bhandari
1 year
We propose a two-step LLM-based approach: 1️⃣ Extract data from reports. 2️⃣ Geotag the data using global admin region databases.
1
0
0
@prb977
Prabin Bhandari
1 year
Epidemiological data is crucial for public health, but extracting and geotagging it from documents is challenging. Our work to be presented at #EMNLP2024’s 3rd NLP4PI Workshop, “From Text to Maps: LLM-Driven Extraction and Geotagging of Epidemiological Data”, tackles this. 🧵
1
2
8
@prb977
Prabin Bhandari
1 year
More Details and analysis in the paper: https://t.co/ovvxGqrz65 Code and Data: https://t.co/YXA0SoBXFx Llama-2-trained model:
Tweet card summary image
huggingface.co
0
0
1
@prb977
Prabin Bhandari
1 year
One key insight? Open source LLMs such as Llama-2, when trained even with a limited amount of actual travel data (Llama-2-trained in the plot), can generate quality synthetic travel surveys to facilitate urban mobility assessment.
1
0
1
@prb977
Prabin Bhandari
1 year
We developed a robust evaluation framework to compare LLM-generated surveys with actual data across three granularities: 1) Pattern-level (overall survey metrics), 2) Trip-level (transition probability norms), and 3) Activity Chain-level (details in the table below).
1
0
1
@prb977
Prabin Bhandari
1 year
Our approach uses LLMs to synthesize urban mobility data, creating scalable, cost-effective insights for smarter urban planning. By generating synthetic travel survey responses, we support comprehensive, data-driven mobility assessments!
1
0
1
@rohanpaul_ai
Rohan Paul
1 year
One of THE largest personas datasets dropped in @huggingface 🤯 21 Million rows and 142GB 🤯 FinePersonas contains detailed personas for creating customized, realistic synthetic data. So now you can integrate unique persona traits into text generation apps. ------- What's a
17
160
1K
@FabrizioRomano
Fabrizio Romano
2 years
Bayer Leverkusen fans, on fire 🔴⚫️🔥
785
14K
168K
@jeremyphoward
Jeremy Howard
2 years
Today, with @Tim_Dettmers, @huggingface, & @mobius_labs, we're releasing FSDP/QLoRA, a new project that lets you efficiently train very large (70b) models on a home computer with consumer gaming GPUs. 1/🧵 https://t.co/UAsWOLtn7a
Tweet card summary image
answer.ai
We’re releasing an open source system, based on FSDP and QLoRA, that can train a 70b model on two 24GB GPUs.
83
644
3K
@FractalEcho
Rua M. Williams
2 years
The racism behind chatGPT that we aren't talking about... This year, I learned that students use chatGPT because they believe it helps them sound more respectable. And I learned that it absolutely does not work. A thread. 🧵
83
4K
24K
@prb977
Prabin Bhandari
2 years
To assess LLMs for geospatial reasoning, we devise an MDS-based experiment to predict a city’s location relative to other cities. Our findings indicate that the model’s performance is within an order of magnitude of what MDS could predict if we had access to actual distance.
0
1
1
@prb977
Prabin Bhandari
2 years
We also prompt for geospatial “awareness”, the perception of space, using geospatial prepositions 'near,' 'close to,' and 'far from,' along with the control word 'and'. LLMs demonstrate an understanding of what “near” or “far” means. See maps visualizing the responses.
1
1
0
@prb977
Prabin Bhandari
2 years
To assess geospatial knowledge encoded, we prompt to predict geo-coordinates of cities. LLMs are better at this task than older models, although predicted values are still quite far off. This discrepancy might be due to lack of enough geospatial data in pre-training datasets.
1
0
0
@prb977
Prabin Bhandari
2 years
Are large language models geospatially knowledgeable? Our SIGSPATIAL paper examines the extent of geospatial knowledge encoded in LLMs, as well as their geospatial awareness and application in reasoning tasks related to geospatial data. Paper:
Tweet card summary image
arxiv.org
Despite the impressive performance of Large Language Models (LLM) for various natural language processing tasks, little is known about their comprehension of geographic data and related ability to...
1
5
10