Some things should be easy but they are not. When I recently wanted to install Ubuntu 18.04 on a 10 year old notebook I was faced with a BIOS that had an early implementation of UEFI that just didn’t play along very nicely with today’s EFI bootloaders. It took me a while to figure this one out and the obvious solution then was to switch-off UEFI in the BIOS and boot with the ‘legacy’ BIOS boot loader that requires GRUB to be put on the MBR (Master Boot Record). So much for the theory. But unfortunately, my Ubuntu 18.04 USB installation stick still wouldn’t boot. After a lot of experimenting I found out why.
Is has become common practice of PC and mobile operating systems to assume that Wifi connectivity means unlimited data volume and is pretty much seen as an invitation to download hundreds of megabytes of software updates. This wrecks havoc on many peoples volume cap when they offer tethered Internet access from one of their mobile devices to other devices while on the road. This really makes me wonder why Google or Apple haven’t yet done anything about this on their mobile devices!?
Network operators in a lot of countries around the world are getting ready at the moment to launch first 5G networks in the 3.x GHz spectrum (i.e. 3GPP band n78). One of the particularly interesting things about this band is that its made for TDD operation and not FDD as most operators are used to. This means the downlink and uplink are on the same channel and separated in time and not on the frequency axis. So how does the network inform the device about when to expect downlink and uplink transmission of the channel in general (not individual reception or transmission opportunities)?
Some things are superseded by other and better things and I find it quite interesting that when this happens, it pretty much goes unnoticed. Let’s take for example 3.5″ disks. If you are old enough to remember, do you remember when you last used one?
Having a lab as a network operator is a great thing. You can test new hardware and software there and once you are happy you can deploy things in the wild and make things better for your subscribers. But every now and then something comes along that you don’t only want to test in the lab but you actually want to have a proper shakedown to test stability and functionality before you let your subscribers use the new technology. So how can this actually be done in practice?
For 3 years, I’ve used a refurbished Lenovo X230 as my main notebook. During those years I’ve changed the disk drive several times, first to a 512 GB SSD, then to a 1 TB SSD and finally to 2 TB SSD. Also, I upgraded the RAM to 16 GB as I make heavy use of Virtual Machines and I even exchanged the display panel when it started to malfunction about a year ago. But the frame is now around 6 years old and the replacement IPS panel from China was starting to show a number of brighter patches that kept growing so it was time to do something about it.
And today another quick post on Eduroam, the federated Internet access authentication solution for students and researchers.
When I recently noticed that one of the certificates in the certificate chain I put together for an Austrian Eduroam user I support would expire early next year, I set out to do some preventive maintenance and put together a new chain. Quite to my surprise it seems to be no longer necessary!
A couple of weeks ago I read the first book of the ‘Themis Files’ trilogy by Sylvain Neuvel and was blown away by the original story. In some cases I find that sequels to an original story might not be bad but have a hard time reaching the same level of storytelling and content of the previous book. So I was wondering how it would be in this case.
Time flies and I can hardly believe that I’ve had a fiber connection at my Paris flat for 4 years now and that fiber is still almost nowhere to be seen in Germany. O.k. I do have vectored VDSL and 100 Mbit/s down, 40 Mbit/s up is not so bad. But fiber is playing in an entirely different league. When I got my fiber Line in Paris four years ago, I measured 264 Mbit/s in the downlink direction and 48 Mbit/s in the uplink direction. So how are things today, i.e. 4 years down the road?
Today I thought I write down some thoughts on a book I noticed when I was at the Living Computer Museum and Labs (LCM+L) in Seattle last year: ‘Atypical Geek Girl’ by Katherine Hitchcock. In her memoirs, Katherine describe her 30 year tech career in IBM that began in the 1960s.