AdaTheDev Profile Banner
Adrian Hills Profile
Adrian Hills

@AdaTheDev

Followers
404
Following
1K
Media
333
Statuses
7K

I write code. I delete code. I build stuff. Often found on the Refactor Tractor. .NET / Azure / AI / Data

Hampshire, UK
Joined February 2009
Don't wanna be here? Send us removal request.
@AdaTheDev
Adrian Hills
4 months
One of my biggest UX gripes with GitHub Copilot (in VS) is that you have to wait for the whole response to be generated before you can read it due to the auto-scrolling as the response streams in. Would be a big plus to be able to stop the auto-scroll!
0
0
0
@AdaTheDev
Adrian Hills
1 year
Anyone else having issues with Azure OpenAI (timing out trying to connect via API, failing to load model deployments in Azure OpenAI Studio...)?
6
0
3
@AdaTheDev
Adrian Hills
2 years
Anyone else having issues getting into the Azure OpenAI Studio today?
0
0
4
@AdaTheDev
Adrian Hills
2 years
Good article (Azure OpenAI) on considerations for right-sizing your Provisioned Throughput deployment, spilling peak workloads over to PAYG for cost optimisation: Right-size your PTU deployment and save big ( https://t.co/p6bU4lZlDr)
0
0
1
@AdaTheDev
Adrian Hills
2 years
Important to make a conscious assessment, else you run the risk of hitting production with a workload that the LLM deployment model/provider cannot fulfil reliably. Depending on use case/customers, you may then be forced into a hasty switch to Provisioned Throughput
0
0
1
@AdaTheDev
Adrian Hills
2 years
PAYG models are great and have low outlay, but usually don't have guarantees and usually higher latency. Provisioned Throughput models give consistency and lower latency but a huge jump up in committed spend, where you end up paying for a lot of quiet time in a non 24/7 scenario
1
0
1
@AdaTheDev
Adrian Hills
2 years
Transitioning to Production with LLM capabilities, when you care about consistency of guaranteed throughput and latency is a key challenge that needs to be considered carefully up front.
1
0
1
@AdaTheDev
Adrian Hills
2 years
Full Fibre 900 - £41.99/month Full Fibre 900 for a customer of many years (me) - £55.54/month Don't know what's going on with @Plusnet these days, but that's shocking. Making me wonder if longer term, it'll be wound down (as part of consolidating Plusnet/BT/EE, per mobile)
0
0
1
@AdaTheDev
Adrian Hills
2 years
Any of my followers (or beyond) have experience with Amazon Bedrock for hosting LLMs (on a token usage basis and/or Provisioned Throughput basis)? Keen to hear thoughts, experiences etc. Please RT!
0
0
1
@AdaTheDev
Adrian Hills
2 years
Early days, but early signs looking promising - around 50% reduction in duration, increased reliability/stability due to stable plan reuse. Even looking like there's potential to use a different (cheaper and faster) LLM when using cached plans.
0
0
1
@AdaTheDev
Adrian Hills
2 years
To save generating plans afresh every time for similar inputs, I'm building a prototype Semantic Kernel "Plan Cache". Still needs LLM (+ Vector Search), but with a much smaller input context, and a more focused task.
1
0
1
@AdaTheDev
Adrian Hills
2 years
Switched over recently to the new gen Planners (Handlebars) as part of Semantic Kernel v1.x upgrade. Next step is to optimise the "plan generation" phase of a workflow. Generating plans is expensive (time and cost) and relatively prone to error.
1
0
1
@AdaTheDev
Adrian Hills
2 years
But the trend is pretty consistent on Azure OpenAI with 1106-Preview. Running tests on isolated model deployments, using low token usage test case (<500 prompt tokens, < 300 completion tokens), various points throughout the day, different machines/code implementations etc etc
0
0
1
@AdaTheDev
Adrian Hills
2 years
Having significant latency issues with GPT 4-Turbo (1106-Preview) on Azure OpenAI, making it consistently, noticeably slower than GPT-4 (0613). Anyone else encountering this? e.g. 48.8s (GPT-4-Turbo) vs. 13.9s (GPT-4) Snapshot times, so fluctuates, but wildly so for 4-Turbo.
1
0
1
@AdaTheDev
Adrian Hills
2 years
Currently watching:
0
0
1
@AdaTheDev
Adrian Hills
2 years
Finishing up prep for an internal talk I'll be giving tomorrow on the Semantic Kernel. There are demos, and then there are Generative-AI demos - let's hope it all behaves!
0
0
1
@AdaTheDev
Adrian Hills
2 years
Enjoyed the keynote, as you'd expect!
0
0
1
@AdaTheDev
Adrian Hills
2 years
Checked in, ready for MS Build AI day to start
0
0
3