Board

Board

2272 bookmarks
Custom sorting
AI Is a Lot of Work
AI Is a Lot of Work
How many humans does it take to make tech seem human? Millions.
·theverge.com·
AI Is a Lot of Work
Tech debt metaphor maximalism
Tech debt metaphor maximalism
I really like the "tech debt" metaphor. A lot of people don't, but I think that's because they either don't extend the metaphor far enough,...
·apenwarr.ca·
Tech debt metaphor maximalism
11 years of hosting a SaaS
11 years of hosting a SaaS
Lots of mistakes, some uptime too.
Use managed services for as long as possible. We did ourselves a big disservice by leaving Heroku after only a few months. We should have stayed on it for years - there was so much time wasted managing servers that could have been done for us during critical early days.
·ghiculescu.substack.com·
11 years of hosting a SaaS
It’s Too Early To Call Winners in AI
It’s Too Early To Call Winners in AI
OpenAI and Microsoft have been crowned as the winners of the generative AI wave. Google is seen as having made severe missteps. Pundits say that incumbents will accrue all the value. There are no moats for startups. But ChatGPT is only 7 months old. Incumbent AI features are working, but we don’t know the counterfactual. New models are being trained and deployed. Open-source models aren’t as good but are catching up or have exceeded proprietary ones in different dimensions (developer tooling, ecosystem, etc.).
·blog.matt-rickard.com·
It’s Too Early To Call Winners in AI
Acting your wage will atrophy your abilities
Acting your wage will atrophy your abilities
Abilities unused will atrophy, so putting in anything less than your best means giving up on what you’re capable of. You can’t save talent or energy for better days, only watch it go to waste. This is a hard truth to accept if you don’t think your company has earned your best. And companies make employees feel like that all the time. T...
·world.hey.com·
Acting your wage will atrophy your abilities
ControlNet QR Codes
ControlNet QR Codes
Image diffusion models like Stable Diffusion can generate all types of images — from photorealistic quality to stylized images (e.g., in the style of Studio Ghibli or Leonardo Da Vinci). We started to control the output of these models with prompt engineering — finding the perfect combination of tokens to mix and match to achieve great images. Then we started curating the dataset — e.g., fine-tuning with images taken with a DSLR camera. Low-rank adaptation (LoRA).
·blog.matt-rickard.com·
ControlNet QR Codes
A Token Efficient Language for LLMs
A Token Efficient Language for LLMs
What would a token-efficient language for LLMs look like? YAML is pervasive as a configuration language partly because of its human readability over JSON. But — what’s more token-efficient if you’re using an LLM? Let’s look at some configuration languages as an example, but the concept also generalizes to programming languages.
·blog.matt-rickard.com·
A Token Efficient Language for LLMs
GameDev.net -- Understanding and Implementing Scene Graphs
GameDev.net -- Understanding and Implementing Scene Graphs
Gamedev.net is the leading resource for game developers, featuring daily news updates, over 1500 featured articles and tutorials, and the most active game development forums anywhere!
·archive.gamedev.net·
GameDev.net -- Understanding and Implementing Scene Graphs
Keep The Main Thing The Main Thing
Keep The Main Thing The Main Thing
The main thing is to keep the main thing the main thing — Stephen Covey An important lesson I’ve learned over the years (and still refining). A few dimensions in my experience. The first is prioritization — actually picking the main thing. It’s often hard at first; there are so many competing priorities. And usually, the initial answer is horizontal scaling rather than vertical scaling — i.e., hire more, allocate more resources, and try more things. But putting some work into prioritizing. Do one thing, and do it well. Ruthlessly cut everything else.
·blog.matt-rickard.com·
Keep The Main Thing The Main Thing
Compounding Optimism
Compounding Optimism
Let me share a little theory I have about optimism, and why progress is so easy to underestimate.
·collabfund.com·
Compounding Optimism
AWS quietly enters the multicloud era
AWS quietly enters the multicloud era
The cloud leader confirmed that two new services introduced Tuesday at re:Invent can be used to manage applications on Microsoft and Google's cloud services.
·protocol.com·
AWS quietly enters the multicloud era
I'm All-In on Server-Side SQLite
I'm All-In on Server-Side SQLite
Ben Johnson has joined Fly.io
As databases evolved, so too did the strategies we use to plug them in to our applications. Almost since Codd, we've divided those apps into tiers. First came the database tier. Later, with memcached and Redis, we got the caching tier. We've got background job tiers and we've got routing tiers and distribution tiers. The tutorials pretend that there are 3 tiers, but we all know it's called "n-tier" because nobody can predict how many tiers we're going to end up with.
If you're careful, using this kind of database can get you a lot of performance. But for general-purpose use, you don't want to run your database off the open headers like a funny car. I thought about the kind of work I'd have to do to make BoltDB viable for more applications, and the conclusion I quickly reached was: that's what SQLite is for.
The most important thing you should understand about Litestream is that it's just SQLite. Your application uses standard SQLite, with whatever your standard SQLite libraries are. We're not parsing your queries or proxying your transactions, or even adding a new library dependency. We're just taking advantage of the journaling and concurrency features SQLite already has, in a tool that runs alongside your application. For the most part, your code can be oblivious to Litestream's existence.
But database optimization has become less important for typical applications. If you have a 1 GB database, an NVMe disk can slurp the whole thing into memory in under a second. As much as I love tuning SQL queries, it's becoming a dying art for most application developers. Even poorly tuned queries can execute in under a second for ordinary databases.
There's a magic number for application latency: responses in 100ms or less feel instantaneous. Snappy applications make happy users. 100ms seems like a lot, but it's easy to carelessly chew it up. The 100ms threshold is so important that people pre-render their pages and post them on CDNs just to reduce latency.
·fly.io·
I'm All-In on Server-Side SQLite
Why I Built Litestream
Why I Built Litestream
Despite an exponential increase in computing power, our applications require more machines than ever because of architectural decisions made 25 years ago. You can eliminate much of your complexity and cost by using SQLite & Litestream for your production applications.
This n-tier architecture sounds simple at first but it has hidden complexity. On a single machine, our server could add an in-memory cache to speed up requests but now data is shared across multiple machines so we must add a memcached or Redis server to share cached data. Database servers are easily overloaded by numerous connections so we have to add intermediate services like PgBouncer to pool connections. If we have events in our system that must be communicated to all our nodes then we need a cluster of Kafka machines.
No, there’s a growing movement of people that see the value of single process applications.
So why is SQLite considered a “toy” database in the application development world and not a production database? That one big issue The biggest problem with using SQLite in production is disaster recovery. If your server dies, so does your data. That’s… not good.
·litestream.io·
Why I Built Litestream
Svelte Origins: A JavaScript Documentary
Svelte Origins: A JavaScript Documentary
🎥 More tech documentaries coming out soon, subscribe to be notified 👉(PS: The 2022 Stack Overflow Developer Survey results just came out and Phoenix overto...
·youtube.com·
Svelte Origins: A JavaScript Documentary
Biography of an Idea
Biography of an Idea
You’ll often find creatives are obsessed with process and tools. A primary tool for me is that I write an awful lot. I write because I realized at art school that you can only draw a small percentage of the attributes of an object. You know, if I were to draw this [holds up a glass], you would understand only 20 percent of its nature. You would have no sense of its weight or material or temperature. You would have no sense of the way that it reacted to its environment. Writing helps me frame the problem. A lot of mistakes are made when you frame a problem, because you could already be dismissing 60 percent, 70 percent of the potential ideas.
·blog.matt-rickard.com·
Biography of an Idea
Understanding GPT tokenizers
Understanding GPT tokenizers
Large language models such as GPT-3/4, LLaMA and PaLM work in terms of tokens. They take text, convert it into tokens (integers), then predict which tokens should come next. Playing …
·simonwillison.net·
Understanding GPT tokenizers
First Impressions of Vision Pro and VisionOS
First Impressions of Vision Pro and VisionOS
In the same way that the introduction of multitouch with the iPhone removed a layer of conceptual abstraction — instead of touching a mouse or trackpad to move an on-screen pointer to an object on screen, you simply touch the object on screen — VisionOS removes a layer of abstraction spatially.
Is it a compelling product, though? It’s a famous Steve Jobs axiom that technology is not enough, that you don’t make compelling products — let alone entire platforms — starting from advanced technology and working backward. You start with a vision for the product and platform experience and then create new technology to make it real.
·daringfireball.net·
First Impressions of Vision Pro and VisionOS
Reduce latency and increase cache hits with Regional Tiered Cache
Reduce latency and increase cache hits with Regional Tiered Cache
Regional Tiered Cache provides an additional layer of caching for Enterprise customers who have a global traffic footprint and want to serve content faster by avoiding network latency when there is a cache miss in a lower-tier, resulting in an upper-tier fetch in a data center located far away
·blog.cloudflare.com·
Reduce latency and increase cache hits with Regional Tiered Cache
When credit cards made airport lounges too much of a good thing
When credit cards made airport lounges too much of a good thing
About five years ago, credit-card issuers opened the features bottle and let out the free-lounge-access genie. From being an exclusive service offered only to premium credit-card holders, free access to airport lounges became the top selling point for every product in a bank’s credit-card pack; it was seen as an aspirational benefit to most. Banks...
·the-ken.com·
When credit cards made airport lounges too much of a good thing