Language Selection

English French German Italian Portuguese Spanish

Kubuntu 6.06

Filed under
Linux
Reviews
-s

With all the Ubuntu excitement passed few days it occurred to me that being a KDE fan moreso than gnome, perhaps Kubuntu might be more my cup of tea. When perusing the downloads it also occurred to me that 'hey I have a 64bit machine now!' So, I downloaded the Kubutu 6.06 desktop amd64 iso. Was it more appealing to a diehard KDE fan? Does 64bit programming make much difference?

The boot is similar to Ubuntu's, in fact it's almost identical except for the more attractive blue coloring and instead of the Ubuntu goldish logo we have the blue kubuntu. Otherwise there didn't seem to be much difference until we reached the splash screen. As attractive as Ubuntu's gui splash might be, kubuntu's is much more so. It's clean and crisp, and I just personally prefer blue.

        

Once you reach the desktop, one finds a blue background that looks like large faint bubbles as a foundation for KDE 3.5.2. It is your basic KDE desktop consisting of kapps for most popular tasks. What's not included on in the iso is installable. Included graphics are Kooka, krita, kpdf, and gwenview. For Internet we find akregator, bluetooth chat, Konqueror, Konversation, kopete, Kppp, krdc, drfb, ktorrent and a wireless lan manager. Multimedia includes amaroK, K3b, Kaffeine, KAudioCreator, kmix, and KsCD.

        

OpenOffice.org 2.02 rounds up the office category as well as KDE deciding kontact is an office application. There are plenty of system tools and utilities as well. There are utils for software packaging, setting alarms, configuring groupware connections, managing your printing jobs, and calculating. System tools consists of Kcron, Keep, KInfocenter, KSysGuard, KSystemLog, Konsole and QTParted.

        

On the desktop as well as in the System menu is an icon for Install. Installing to the harddrive is simplified over comparable Linux system, and in this case it is very similar if not identical to the process found in Ubuntu. It starts with answering a few configuration questions such as language, timezone, keyboard, and user and machine name.

        

Next comes partitioning if necessary and setting the target partition and swap. Confirm settings and press Install. All one does now is wait. It takes about 10 minutes for the installer to complete it's work before asking if you'd like to reboot. That's it.

        

Like Ubuntu, the installer presumes you would like grub installed so doesn't bother to ask and my first install attempt wasn't successful. The newly installed system would not boot. It just sat at the 'loading grub' screen blinking at me, in much the same manner as I encountered with the Ubuntu release candidate. After replacing grub with lilo, kubuntu tried to boot, but lots of things failed including the loading of needed modules and the start of the gui. I booted the livecd and tried again, this time doing nothing else in the background and achieved a bootable install. The first time I was taking a bunch of screenshots. I think I'm beginning to see a pattern emerge here in all my installs of the Ubuntu family and can sum it up in a few words of advice. Do not do anything else while your new Ubuntu system installs. This of course detracts from the main advantage of using a livecd as an install medium, but on the other hand, it takes such a short span of time to install that it's not a major sacrifice.

The installed system affords one the opportunity to install whatever applications one might need as well any 3rd party or proprietary drivers. (k)ubuntu software is installed thru an app called adept. Not only is it an software manager, but it also takes care of system or security updates. In fact one of the first thing I saw when I booted Kubuntu the first time was an icon in the System tray for adept and clicking on it brought up an updater. Click to fetch list of updates and in a few seconds it will inform you if anything needs updating. In this case there were updates to the adept software manager and gnome install data. One can Apply Updates or Forget Changes and Quit. I clicked Apply Changes and the updates were downloaded and installed in seconds without issue.

        

In the menu is an entry for Adept which opens a window similar to Synaptic. You can search for specific packages by keywork with tickable options, and right click package name to "Request Install." Then click on the Apply Changes button and your package as well as dependencies are downloaded and installed.

        

Clicking on "Add and Remove Programs" also brings up adept, but in a different layout. In this layout one finds the applications available or installed listed by category. Ticking the little checkbox and clicking Apply Changes will install or remove chosen programs.

        

The hardware detection was good and pretty much everything worked out of the box. Kaffeine was able to play mpgs and the example files but not avis. OpenOffice crashed and disappeared my first attempt at using it, but functioned properly in all subsequent tests. The KDE that was included was a bit stripped down and included no games at all, but lots of choices are available through the software manager. The desktop itself was pretty even if customized very little. Under the hood is a 2.6.15 kernel, Xorg 7.0 and gcc 4.0.3 is installable.

The performance of the system was well above average. In fact, I'll just say it, that thing flies. Applications opened up before I could move my mouse. There was no artifacting or delay in redrawing windows, no delay at all in switching between windows, or "jerkiness" when moving windows around. The menu popped right open without delay as well. The whole system felt light and nimble. I was quite impressed. Comparing the performance of kde kubuntu to gnome ubuntu is almost like comparing peaches to nectarines and since I didn't test the x86 version of kubuntu, I can't say with any authority or expertise that kubuntu 64 out-performs the others. But I can say this is one of the, if not the, fastest full-sized systems I've tested. Yes sir, kubuntu was quite impressive.

Kubuntu 6.06 Review

Thanks for your review.
Just a few personal comments about this new release.

FIRST, THE BAD . . .

1) Yes, I profoundly dislike NOT being able to log in as root. Granted, I don't spend much time logged-in as user root, but when I'm going to do an extended session of system configuration and maintenance, it's the fastest, most efficient way. And, of course, Linux is supposed to be about choice.

So I googled "Kubuntu root login", and got to a Kubuntu forum where some user had asked the same question. In the forum, the next person had posted in reply (I'm paraphrasing here): I know how to enable root logins for Kubuntu, but I'm not going to tell you how because this is not a good idea.

I just couldn't believe my eyes. OSS is all about freedom and the ability to control your own machine. "THERE'S NO INFORMATION HIDING IN LINUX!!!" (Imagine Tom Hanks in the movie, A League of Their Own saying "There's no crying in baseball!")

As I sat there thinking about this, I began to think that these Ubuntu/Kubuntu folk are a different kind of folk than me--alien, strange, ungracious, and infuriating. All that wondering about the popularity of Ubuntu/Kubuntu increased. Why would anyone want to use a distro where a simple request for information was received with such an ignorant, uptight, shortsighted, narrowminded, hypocritical, and outright anal response?

Working myself up to a good simmer, now, I then look at the next post in the forum, where a user quietly and considerately told exactly how to do it. OK, maybe these Kubuntu folk aren't jerks. Too much rush to judgement and stereotyping going on in our world anyway.

2) I'm not too familiar with Debian, or Debian based distros--so this one is probably my fault. I couldn't get Nvidia's accelerated drivers working properly. I have an older 17" LCD monitor that still works great, but it's very fussy about sync rates. I can usually tinker around and get things working, but no go here. I admit I was impatient here, and if I'd spent more time, I could have gotten it to work.

3) Development compilers and libraries are not included with the basic live CD install. There is such a thing as designing a distro for beginners, but not treating them like idiots. Can't find a package for your distro that works? Then, you can get the source and compile your own. Basic tools to do this, in my opinion, should always be included with the basic install of a Linux distro.

NOW, THE GOOD.

1) I like the graphical package installer, Adept. It works well, and is well designed and integrated.

2) Despite the heavy load the Ubuntu/Kubuntu servers must have been undergoing with a new release, they were quick and responsive.

3) I agree with every thing srlinux had to say in her (typically excellent) review, particularly when she says Kubuntu is very fast and responsive. I installed the amd64 version, and speed was excellent.

4) Kubuntu very courteously found all the other distros on my hard disks, and added them to the Grub boot menu. If all distros would do this, installing and trying out multiple distros would certainly be easier.

Well, that's it. I think it might be interesting to see an in-depth comparison of the latest Mepis(ver. 6, rc4) to this new Kubuntu release.

Gary Frankenbery

re: Kubuntu 6.06 Review

gfranken wrote:

Thanks for your review.
Just a few personal comments about this new release.

FIRST, THE BAD . . .

1) Yes, I profoundly dislike NOT being able to log in as root. Granted, I don't spend much time logged-in as user root, but when I'm going to do an extended session of system configuration and maintenance, it's the fastest, most efficient way. And, of course, Linux is supposed to be about choice.

Yeah, I didn't like that one too much, especially when the ubuntus hit the pipes. I'm used to it now I guess and it only annoys me slightly when I forget to sudo a command. I found just setting a root password will fix that tho.

teehee on the tom hanks thing. Big Grin

gfranken wrote:

3) Development compilers and libraries are not included with the basic live CD install. There is such a thing as designing a distro for beginners, but not treating them like idiots. Can't find a package for your distro that works? Then, you can get the source and compile your own. Basic tools to do this, in my opinion, should always be included with the basic install of a Linux distro.

Yeah, that used to really urk me too. But like the sudo thing, I'm getting used to it. I'm seeing it more and more in distros these days. A comment on another site said to install the build-essentials to get gcc, make, and friends. It was nice seeing that comment before my installs that way I didn't have to get all annoyed at it. Big Grin

Thanks for your kind words for me, I really appreciate that.

----
You talk the talk, but do you waddle the waddle?

It's not as bad as it looks

1) Yes, I profoundly dislike NOT being able to log in as root.

I must admit that I created a root password on the first machines where I installed Ubuntu. (sudo passwd) But now I just let the root user blocked and use (sudo su) to get a shell that has root rights and exit it again after some system maintenance. It is possible more secure that script kiddies cannot login as root via ssh (not standard installed but the first thing I usually install) even if they try really hard.

2) I couldn't get Nvidia's accelerated drivers working properly.

After installing the nvidia package (apt-get install nvidia-glx) I usually run the reconfiguration of xorg by (dpkg-reconfigure xserver-xorg) this creates a new /etc/X11/xorg.conf file including sync ranges. This tool always shows the options chosen the last time it has run so it is easy to use multiple times to tweak things.

3) Development compilers and libraries are not included with the basic live CD install.

It is easy to correct (apt-get install build-essential) but many people don't need those tools. The current repositories are packed with all kind of usefull software. It is very seldom that I revert to debian-unstable to get sources of something and repack it to create an ubuntu version of the same project.

Re: It's not as bad as it looks

Quote:
After installing the nvidia package (apt-get install nvidia-glx) I usually run the reconfiguration of xorg by (dpkg-reconfigure xserver-xorg) this creates a new /etc/X11/xorg.conf file including sync ranges. This tool always shows the options chosen the last time it has run so it is easy to use multiple times to tweak things.
Thanks Jurgen, for taking the time to explain the nvidia thing. My weakness when configuring Debian based distros is definitely showing here.

Also thanks for your clarification on installing the development packages.

Kubuntu is definitely on my radar. It was certainly blazingly fast.

Regards,
Gary Frankenbery
Computer Science Teacher
Grants Pass High School, Oregon, USA

Comment viewing options

Select your preferred way to display the comments and click "Save settings" to activate your changes.

More in Tux Machines

today's howtos and leftovers

  • Linux commands for user management
  • CONSOOM All Your PODCASTS From Your Terminal With Castero
  • Install Blender 3D on Debian 10 (Buster)
  • Things To Do After Installing openSUSE Leap 15.2
  • GSoC Reports: Fuzzing Rumpkernel Syscalls, Part 2

    I have been working on Fuzzing Rumpkernel Syscalls. This blogpost details the work I have done during my second coding period.

  • Holger Levsen: DebConf7

    DebConf7 was also special because it had a very special night venue, which was in an ex-church in a rather normal building, operated as sort of community center or some such, while the old church interior was still very much visible as in everything new was build around the old stuff. And while the night venue was cool, it also ment we (video team) had no access to our machines over night (or for much of the evening), because we had to leave the university over night and the networking situation didn't allow remote access with the bandwidth needed to do anything video. The night venue had some very simple house rules, like don't rearrange stuff, don't break stuff, don't fix stuff and just a few little more and of course we broke them in the best possible way: Toresbe with the help of people I don't remember fixed the organ, which was broken for decades. And so the house sounded in some very nice new old tune and I think everybody was happy we broke that rule.

Programming Leftovers

  • Podcast: COBOL development on the mainframe

    Nic reached out when COBOL hit the news this spring to get some background on what COBOL is good for historically, and where it lives in the modern infrastructure stack. I was able to talk about the basics of COBOL and the COBOL standard, strengths today in concert with the latest mainframes, and how COBOL back-end code is now being integrated into front ends via intermediary databases and data-interchange formats like JSON, which COBOL natively supports.

  • What I learned while teaching C programming on YouTube

    The act of breaking something down in order to teach it to others can be a great way to reacquaint yourself with some old concepts and, in many cases, gain new insights. I have a YouTube channel where I demonstrate FreeDOS programs and show off classic DOS applications and games. The channel has a small following, so I tend to explore the topics directly suggested by my audience. When several subscribers asked if I could do more videos about programming, I decided to launch a new video series to teach C programming. I learned a lot from teaching C, and in the process, I came across some meaningful takeaways I think others will appreciate. Make a plan For my day job, I lead training and workshops to help new and emerging IT leaders develop new skills. Outside of regular work, I also enjoy teaching as an adjunct professor. So I'm very comfortable constructing a course outline and designing a curriculum. That's where I started. If you want to teach a subject effectively, you can't just wing it. Start by writing an outline of what topics you want to cover and figure out how each new topic will build on the previous ones. The "building block" method of adding new knowledge is key to an effective training program.

  • Google's Flutter 1.20 framework is out: VS Code extension and mobile autofill support
  • Google Engineers Propose "Machine Function Splitter" For Faster Performance

    Google engineers have been working on the Machine Function Splitter as their means of making binaries up to a few percent faster thanks to this compiler-based approach. They are now seeking to upstream the Machine Function Splitter into LLVM. The Machine Function Splitter is a code generation optimization pass for splitting code functions into hot and cold parts. They are doing this stemming from research that in roughly half of code functions that more than 50% of the code bytes are never executed but generally loaded into the CPU's data cache.

  • Modernize network function development with this Rust-based framework

    The world of networking has undergone monumental shifts over the past decade, particularly in the ongoing move from specialized hardware into software defined network functions (NFV) for data plane1 and packet processing. While the transition to software has fashioned the rise of SDN (Software-defined networking) and programmable networks, new challenges have arisen in making these functions flexible, efficient, easier to use, and fast (i.e. little to no performance overhead). Our team at Comcast wanted to both leverage what the network does best, especially with regards to its transport capacity and routing mechanisms, while also being able to develop network programs through a modern software lens—stressing testing, swift iteration, and deployment. So, with these goals in mind, we developed Capsule, a new framework for network function development, written in Rust, inspired by Berkeley's NetBricks research, and built-on Intel's Data Plane Development Kit (DPDK).

  • This Week in Rust 350
  • Firefox extended tracking protection

    This Mozilla Security Blog entry describes the new redirect-tracking protections soon to be provided by the Firefox browser.

  • Karl Dubost: Browser developer tools timeline

    I was reading In a Land Before Dev Tools by Amber, and I thought, Oh here missing in the history the beautifully chiseled Opera Dragonfly and F12 for Internet Explorer. So let's see what are all the things I myself didn't know.

  • Daniel Stenberg: Upcoming Webinar: curl: How to Make Your First Code Contribution

    Abstract: curl is a wildly popular and well-used open source tool and library, and is the result of more than 2,200 named contributors helping out. Over 800 individuals wrote at least one commit so far. In this presentation, curl’s lead developer Daniel Stenberg talks about how any developer can proceed in order to get their first code contribution submitted and ultimately landed in the curl git repository. Approach to code and commits, style, editing, pull-requests, using github etc. After you’ve seen this, you’ll know how to easily submit your improvement to curl and potentially end up running in ten billion installations world-wide.

Security: Zoom Holes, New Patches and etcd Project Security Committee

  • Zoombomber crashes court hearing on Twitter hack with Pornhub video
  • Security updates for Wednesday

    Security updates have been issued by Debian (net-snmp), Fedora (mingw-curl), openSUSE (firefox, ghostscript, and opera), Oracle (libvncserver and postgresql-jdbc), Scientific Linux (postgresql-jdbc), SUSE (firefox, kernel, libX11, xen, and xorg-x11-libX11), and Ubuntu (apport, grub2, grub2-signed, libssh, libvirt, mysql-8.0, ppp, tomcat8, and whoopsie).

  • The CNCF etcd project reaches a significant milestone with completion of security audit

    This week, a third-party security audit was published on etcd, the open source distributed key-value store that plays a crucial role in scaling Kubernetes in the cloud. For etcd, this audit was important in multiple ways. The audit validates the project’s maturity and sheds light on some areas where the project can improve. This sort of audit is required criteria for any project in the Cloud Native Computing Foundation (CNCF) to qualify for graduation from the CNCF. Read the CNCF blog post that I co-authored to learn more about the audit and what it uncovered. As one of the project maintainers and one of two members of the etcd Project Security Committee, I’d love to share a few reasons I’m hopeful for etcd’s future and why now is a great time to contribute to etcd’s open source community.

Linux Plumbers Conference and Kernel Developments in METRICFS, FS-Cache, HWMON

  • Application Ecosystem Microconference Accepted into 2020 Linux Plumbers Conference

    We are pleased to announce that the Application Ecosystem Microconference has been accepted into the 2020 Linux Plumbers Conference! The Linux kernel is the foundation of the Linux systems, but it is not much use without applications that run on top of it. The application experience relies on the kernel for performance, stability and responsiveness. Plumbers is the perfect venue to have the kernel and app ecosystems under one roof to discuss and learn together and make a better application experience on the Linux platform.

  • Google Opens Patches For "METRICFS" That They Have Used Since 2012 For Telemetry Data

    The METRICFS file-system has been in use internally at Google since 2012 for exporting system statistics to their telemetry systems with around 200 statistics being exported per machine. They are now posting the METRICFS patches as open-source for review and possible upstreaming. A "request for comments" on METRICFS was sent out today on the Linux kernel mailing list. Their motives for now finally publishing these patches is as a result of the recent Statsfs proposal by a Red Hat engineer for a RAM-based file-system for exposing kernel statistics to user-space. METRICFS has a similar aim to Statsfs.

  • FS-Cache Rewritten But Even Its Developers Are Hesitant About Landing It For Linux 5.9

    FS-Cache provides the Linux kernel with a general purpose cache for network file-systems like NFS and AFS but also other special use-cases like ISO9660 file-systems. FS-Cache has been rewritten for better performance and reliability, among other benefits, and while it has been sent in as a pull request for Linux 5.9 even its own developers provide some caution over landing it this cycle. FS-Cache has seen work to "massively overhaul" it with a variety of improvements. The new and improved FS-Cache will now use async direct I/O in place of snooping for updated pages that in turn means less virtual memory overhead. The new FS-Cache implementation has simpler object management, changes to object invalidation, and a variety of other work.

  • Corsair Commander Pro Driver Sent In To Linux 5.9

    The hardware monitoring (HWMON) subsystem has a new driver that is likely to excite some enthusiasts wanting greater control over thermal monitoring and fan control for their systems. The previously covered Corsair Commander Pro Linux driver is now coming with Linux 5.9. The Commander Pro offers six 4-pin fan ports with PWM controls, two RGB LED channels, and four thermal sensors. An interested user/developer created this Linux driver without the support from Corsair. The thermal and fan control support is in place with this new HWMON driver while the RGB lighting controls are available from OpenRGB.