Some lighter "non-hype" reflections
Again, I posted this on LinkedIn and wanted to share it here as well.
Every year I go to North America for a month or so in Feb because of conferences.
I also take this time out to reflect on the previous year, take a step back from my life in the UK and recalibrate.
This time I was in beautiful British Columbia šØš¦, and once I kept to myself like Tony Stark in a bunker, I took the opportunity to play, build and revisit things Iād bookmarked over the past few months.
Here are some of my lighter, non-hype (and maybe slightly controversial) thoughts during that time away:
āVibe codedā websites can look good⦠when you know what youāre doing. Vibe coding does pose its problems, but for the purpose of this post, Iāll put that aside. Iāve been working on my own websites (using Claude Code) and itās clear that model outputs have improved over the past year. Designs are more aesthetically pleasing, although without strong direction they likely converge towards the same design patterns which probably explains why so many sites look similar.
āOpen-sourceā AI still isnāt really open-source. Youād think by now weād have an influx of fully open models, and while there are a few, what we mostly have are open-weight models, which isnāt quite the same thing.
Small Language Models are being slept on tbh. On-device, lightweight AI just makes sense for a lot of practical use cases.
Most static LLM benchmarks are useless. This isnāt a new discovery, but itās becoming clearer that real-world evaluation is what truly matters. I like what the team at Arena. AI are doing here, using community comparisons and feedback to build leaderboards for real-world use....check them out.
Consumer āself-drivingā cars are starting to feel more real. I rented a Tesla Model 3 with FSD for a month (FSD's not available in the UK) and was genuinely surprised at how well it handled real-world driving. Curious to see how BYDās "Godās Eye" tech compares.
AGI is still a useless term. Itās ill-defined and everyone means something different by it. Itās also become a kind of red herring, distracting folks from the more immediate concerns.
Itās become increasingly clear that scaling LLMs alone isnāt sufficient in the quest for general intelligence (I know, I know š¬) and so we now have alternative approaches (world model approaches JEPA etc.) gaining attention (and funding).
There is nothing groundbreaking in this post, but merely a few non-hype observations from my time away.
(Pictured below with my cousin Neeta Nagra Ed.D, MBA and her dog Koko, at Tunnels Bluff, Lionās Bay, BC šØš¦. Awesome hike, much recommended, the pics donāt do it justice).



