Skip to main content

Linux plunge not working out so well

So I don't think Linux is ready for laptops. Well, my laptop anyway.

The first problem I encountered was with screen resolution. I didn't have an option for the "native" 1280x800 widescreen resolution of my Dell 700m. After digging through the Ubuntu support forums, I discovered that I had to install a small utility called 915resolution. It was a minor pain to track this down, as there were several contradictory sets of instructions found with Google, but running this command:
sudo apt-get install 915resolution
and restarting seemed to fix things.

My next problem was with WiFi. Ubuntu's network management applet didn't show any available wireless networks, despite the fact that I know there are dozens nearby my home. Reboot into Windows, do some more browsing, and discover some diagnostic tests to run. Boot back into Ubunutu. It appears that command-line tools can see wireless networks nearby, but Ubuntu's GUI is broken and doesn't list them. I could deal with having to run a few commands to connect, but...

It also seems that Ubuntu does not support Wifi Protected Access 2 (WPA2) wireless access points out of the box. This is a much bigger deal, as my home network is all WPA2 WiFi, and I will not "downgrade" to any version of WEP, as it is woefully insecure.

So I boot back into Windows, and do a few more hours of research. There are WPA tools for Debian-derived Linux systems, and some folks have gotten them to work. But I was really unwilling to go down this route, as the documented procedures were pages long, and involved running a command line utility to generate a password hash for each new network I wanted to use. Not exactly useful for someone who needs to do work on the go.

Finally, I looked into getting support for my Sprint Mobile Broadband Card, which provides about 1 Mbps download speed just about anywhere. This little device is my lifeline for work. From what I read on the net (again scattered over dozens of contradictory sites), there is almost no driver support at all for these mobile Wireless cards in Linux. To get something working, I would have to modify some available driver for another device and compile it into my kernel.

So I gave up. I have a family, and a job, and I just wanted to get some work done. Right now, at least, Ubuntu doesn't have enough mobile device support for my needs.

And I know all the Linux fanboys out there will call me an 1d10t n00b, and blame the hardware manufacturers for not releasing good open-source drivers. But you know what? I don't care. Ubuntu failed me. Going mobile with Windows XP is light-years easier by comparison, and I'm not going to switch to something that requires so much manual configuration each time I want to work on the road.

Maybe I'll try a MacBook instead.


zuttobenkyou said…
I've had similar experiences with Ubuntu 7.04 and WiFi on my Dell Latitude D505.

I'm currently waiting for version 8 to come out for Ubuntu this spring. It will hopefully fix some problems relating to multi-monitor support (last time I checked, Ubuntu's support for 2 LCD monitors was pretty much nonexistant).
RPM said…
zuttobenkyou, I have recently upgraded to a Dell D420, but it has the same graphics chipset, and same WiFI chipset, so I think I too will wait for Ubuntu 8. I will also try a Fedora bootable CD to see if it sees all the netowrk drivers. The lack of reasonable support for the mobile broadband cards is still a killer for me. Yes, I know this is "the vendor's fault", but as far as I can tell no U.S. mobile carrier has Linux support for their broadband cards.

Popular posts from this blog

Fixing slow NFS performance between VMware and Windows 2008 R2

I've seen hundreds of reports of slow NFS performance between VMware ESX/ESXi and Windows Server 2008 (with or without R2) out there on the internet, mixed in with a few reports of it performing fabulously.
We use the storage on our big Windows file servers periodically for one-off dev/test VMware virutal machines, and have  been struggling with this quite a bit recently. It used to be fast. Now it was very slow, like less than 3 MB/s for a copy of a VMDK. It made no sense.
We chased a lot of ideas. Started with the Windows and WMware logs of course, but nothing significant showed up. The Windows Server performance counters showed low CPU utilization and queue depth, low disk queue depth, less than 1 ms average IO service time, and a paltry 30 Mbps network utilization on bonded GbE links.
So where was the bottleneck? I ran across this Microsoft article about slow NFS performance when user name mapping wasn't set up, but it only seemed to apply to Windows 2003. Surely the patch me…

Google's public NTP servers?

I was struggling with finding a good set of low-ping NTP servers for use as upstream sources in the office. Using is great and all, but the rotating DNS entries aren't fabulous for Windows NTP clients (or really any NTP software except the reference ntpd implementation).

ntpd resolves a server hostname to an IP once at startup, and then sticks with that IP forever. Most other NTP clients honor DNS TTLs, and will follow the rotation of addresses returned by This means Windows NTP client using the built-in Windows Time Service will actually be trying to sync to a moving set of target servers when pointed at a source. Fine for most client, but not great for servers trying to maintain stable timing for security and logging purposes.

I stumbled across this link referencing Google's ntp servers at hostname time[1-4] These servers support IPv4 and IPv6, and seem to be anycast just like Google's public DNS servers at time…

Presets versus quality in x264 encoding

I'm scoping a project that will require re-encoding a large training video library into HTML5 and Flash-compatible formats. As of today, this means using H.264-based video for best compatability and quality (although WebM might become an option in a year or two).
The open source x264 is widely considered the state of the art in H.264 encoders. Given the large amount of source video we need to convert as part of the project, finding the optimal trade-off between encoding speed and quality with x264-based encoders (x264 itself, FFmpeg, MEencoder, HandBrake, etc.) is important.
So I created a 720p video comprised of several popular video test sequences concatenated together. All of these sequences are from lossless original sources, so we are not re-compressing the artifacts of another video codec. The sequences are designed to torture video codecs: scenes include splashing water, flames, slow pans, detailed backgrounds and fast motion. I did several two-pass 2500 kbps encodings using …