Any recs for screencasts of people building with Claude Code/Cursor/Gemini CLI/etc?
Doesn't have to be anything complex or long, just clearly showing workflow/interaction/etc.
I like computers, Korean, and computers-and-Korean.
Language and keyboard stuff @Google and PhD student in Tokyo.
Georgia Tech -> 연세대학교 -> 東京工業大学.
Interested in high school CS education.
Academic:
- #ML
- #NLProc
- #Automata
- #CJK
- #Tokenization
- #FederatedLearning
- #Compression
Programming:
- #JuliaLang
- #Python
- #OSS
Other:
- High School CS #Education
- #Korean
- #Keyboards
- #Esperanto
- #Chess
- #Coffee
Any recs for screencasts of people building with Claude Code/Cursor/Gemini CLI/etc?
Doesn't have to be anything complex or long, just clearly showing workflow/interaction/etc.
@pbloem In some sense, I agree that character-level would solve a lot of our problems. But character-level models have low information density and, at least with current architectures, are too costly and slow.
Subword tokenization is definitely an imperfect solution, but improves upon both of those problems for most (but ofc then has the downsides of like "has trouble spelling/doing basic math", etc.).
I hope to see a lot more research on tokenization-free methods like BLT.
@pbloem Here you go!
@drgroftehauge yep, I guess what I normally do is not quite code golf though. I try to find one-liners, but they don't necessarily have the fewest number of characters (the real goal of golfing).
This problem came up again, so I updated my old solution!
A few weeks ago I got bored and tried solving a leetcode problem "at random". Basically, I wanted to find a way to set a random seed such that randomly generating an answer solved all the test cases for the following problem: https://leetcode.com/problems/find-unique-binary-string/.
Through a little trial and error, I managed to solve it. Check out the writeup!
https://mcognetta.github.io/posts/leetcode-random-seed/
#python #leetcode #codegolf #programming #MaliciousCompliance
Today we are launching a server dedicated to Tokenization research! Come join us!
discord.gg/CDJhnSvU
!!Con 2024 is just three weeks away! Join us for two days of ten-minute talks about the joy, excitement, and surprise of computing!
🎟️ Get tickets and learn more: https://bangbangcon.com
@sharif Ah, for keystroke-level golf I agree. But I prefer to do line-level golf, where I pack everything into a giant one liner (so ";" is disallowed). It's a lot of fun.
There is a workaround (that I don't particularly love) for this where you can get one line:
```
(x := [1,2,3], [z for z in x])[1]
```
You can also usually work around it with map/other functional stuff, but it would be so much nicer if I could just do it with a walrus operator.
It's frustrating that you can't use the walrus operator in list comprehensions where you can in the unrolled loop.
Not that I would do it often in real code, it's just annoying for when I want to golf.
There has been a remarkable breakthrough towards the Riemann hypothesis (though still very far from fully resolving this conjecture) by Guth and Maynard making the first substantial improvement to a classical 1940 bound of Ingham regarding the zeroes of the Riemann zeta function (and more generally, controlling the large values of various Dirichlet series): https://arxiv.org/abs/2405.20552
Let 𝑁(σ,𝑇) denote the number of zeroes of the Riemann zeta function with real part at least σ and imaginary part at most 𝑇 in magnitude. The Riemann hypothesis tells us that 𝑁(σ,𝑇) vanishes for any σ>1/2. We of course can't prove this unconditionally. But as the next best thing, we can prove zero density estimates, which are non-trivial upper bounds on 𝑁(σ,𝑇). It turns out that the value σ=3/4 is a key value. In 1940, Ingham obtained the bound \(N(3/4,T) \ll T^{3/5+o(1)}\). Over the next eighty years, the only improvement to this bound has been small refinements to the 𝑜(1) error. This has limited us from doing many things in analytic number theory: for instance, to get a good prime number theorem in almost all short intervals of the form \((x,x+x^\theta)\), we have long been limited to the range \(\theta>1/6\), with the main obstacle being the lack of improvement to the Ingham bound. (1/3)
I can't be stopped
I spent a bit of time last night and this morning over-optimizing a naive #Python #LeetCode solution to get the fastest solution on the site.
Enjoy: https://theoreticallygoodwithcomputers.com/posts/leetcode-gold-optimization/
Starting in 5 minutes!