• Shortcuts : 'n' next unread feed - 'p' previous unread feed • Styles : 1 2

» Publishers, Monetize your RSS feeds with FeedShow:  More infos  (Show/Hide Ads)


Date: Thursday, 18 Sep 2014 05:04

The new Ubuntu Touch operating system from Canonical will power the new Meizu MX4 phone and it will be out in December, according to the latest information posted by the Chinese company. We now take a closer look at this new phone to see how it will hold up with an Ubuntu experience.

Canonical hasn’t provided any kind of information about a timetable for the launch of the new Ubuntu phone from Meizu, and even the information that we have right now has been posted initially on an Italian blog of the Chinese company. Basically, no one is saying anything officially, but that’s not really the point.

The new Meizu MX4 was announced just a couple of weeks ago and many Ubuntu users have asked themselves if this is the phone that will eventually feature the upcoming Ubuntu Touch. It looks like that is the case, so we now take a closer look at this powerful handset.

Source:

http://news.softpedia.com/news/Everything-You-Need-to-Know-About-Meizu-MX4-the-Upcoming-Ubuntu-Phone-458882.shtml

Submitted by: Silviu Stahie

Author: "kr1st0"
Send by mail Print  Save  Delicious 
Date: Thursday, 18 Sep 2014 00:28

In order to have more productivity under my environment, as a command line centric guy, I started three years ago to use zsh as my defaul shell. And for who never tried it, I would like to share my personal thoughts.

What are the main advantages?

  • Extended globbing: For example, (.) matches only regular files, not directories, whereas az(/) matches directories whose names start with a and end with z. There are a bunch of other things;
  • Inline glob expansion: For example, type rm *.pdf and then hit tab. The glob *.pdf will expand inline into the list of .pdf files, which means you can change the result of the expansion, perhaps by removing from the command the name of one particular file you don’t want to rm;
  • Interactive path expansion: Type cd /u/l/b and hit tab. If there is only one existing path each of whose components starts with the specified letters (that is, if only one path matches /u/l/b*), then it expands in place. If there are two, say /usr/local/bin and /usr/libexec/bootlog.d, then it expands to /usr/l/b and places the cursor after the l. Type o, hit tab again, and you get /usr/local/bin;
  • Nice prompt configuration options: For example, my prompt is currently displayed as tov@zyzzx:/..cts/research/alms/talk. I prefer to see a suffix of my current working directory rather than have a really long prompt, so I have zsh abbreviate that portion of my prompt at a maximum length.

Font: http://www.quora.com/What-are-the-advantages-and-disadvantages-of-using-zsh-instead-of-bash-or-other-shells

The Z shell is mainly praised for its interactive use, the prompts are more versatilly, the completion is more customizable and often faster than bash-completion. And, easy to make plugins. One of my favorite integrations is with git to have better visibility of current repository status.

As it focus on interactive use, is a good idea to keep maintaining your shell scripts starting with #!/bin/bash for interoperability reasons. Bash is still most mature and stable for shell scripting in my point of view.

So, how to install and set up?

sudo apt-get install zsh zsh-lovers -y

zsh-lovers will provide to you a bunch of examples to help you understand better ways to use your shell.

To set zsh as the default shell for your user:

chsh -s /bin/zsh

Don't try to set zsh as default shell to your full system or some things should stop to work.

Two friends of mine, Yuri Albuquerque and Demetrius Albuquerque (brothers of a former hacker family =x) also recommended to use https://github.com/robbyrussell/oh-my-zsh. Thanks for the tip.

How to install oh-my-zsh as a normal user?

curl -L http://install.ohmyz.sh | sh

My $ZSH_THEME is seted to "bureau" under my $HOME/.zshrc. You can try "random" or other themes located inside $HOME/.oh-my-zsh/themes.

For command-not-found integration:

echo "source /etc/zsh_command_not_found" >> ~/.zshrc

If you doesn't have command-not-found package:

sudo apt-get install command-not-found -y

And, if you use Ruby under RVM, I also recommend to read this:
http://rvm.io/integration/zsh

Happy hacking :-)

Author: "Ayrton Araújo"
Send by mail Print  Save  Delicious 
Date: Wednesday, 17 Sep 2014 13:11

After Remy “Remington” Sharp and Bruce “Bruce” Lawson published Introducing HTML5 in 2010, the web development community have been eager to see what they’ll turn their talents to next.1 Now their new book is out, Responsive Design for Dummies.

It’s… got its good points and its bad points. As the cover proudly proclaims, they fully embrace the New World Order of delivering essential features via Web Components. I particularly liked their demonstration of how to wrap a whole site inside a component, thus making your served HTML just be <bruces-site> and so saving you bandwidth2. Their recommendation that Flickr and Facebook use this approach to stop users stealing images may be the best suggestion for future-proofing the web that we’ve heard in 2014 so far. The sidebar on how to use this approach and hash-bang JavaScript URLs together ought to become the new way that we build everything, and I’m eager to see libraries designed for slow connections and accesssibility such as Angular.js adopt the technique.

Similarly, the discussion of how Service Workers can deliver business advantages on the Apple iWatch was welcome, particularly given the newness of the release. It’s rare to see a book this up-to-date and this ready to engage with driving the web forward. Did Bruce and Remy get early access to iWatch prototypes or something? I am eager to start leveraging these techniques with my new startup3.

It’s not all perfect, though. I think that devoting three whole chapters to a Dawkins-esque hymn of hatred for everyone who opposed the <picture> element was a bit more tactless than I was hoping for. You won, chaps, there’s no need to rub it in.4

I’d also like to see, if I’m honest, ideas for when breakpoints are less appropriate. I appreciate that the book comes with a free $500 voucher for Getty Images, but after at Bruce and Remy’s recommendation I downloaded separate images for breakpoints at 17px, 48px, 160px, 320px, 341px, 600px, 601px, 603px, 631px, 800px, 850px, 900px, 1280px, 2560px, and 4200px for retina Firefox OS devices, I only had $2.17 left to spend and my server has run out of disc space. Even after using their Haskell utility to convert the images to BMP and JPEG2000 formats I still only score 13.6% on the Google Pagespeed test, and my router has melted. Do better next time, chaps.

Nonetheless, despite these minor flaws, and obvious copy-editing flubs such as “responsive” being misspelled on the cover itself5, I’d recommend this book. Disclaimer: I know both the authors biblicallypersonally and while Bruce has indeed promised me “a night to remember” for a positive review, that has not affected at all my judgement of this book as the most important and seminal work in the Web field since Kierkegaard’s “Sarissa.js Tips and Tricks”.

Go and buy it. It’s so popular that it might actually be hard to find a copy, but if your bookseller doesn’t have it, you should shout at them.

  1. other than inappropriate swimwear, obviously
  2. I also liked their use of VML and HTML+TIME in a component
  3. it’s basically Uber for pie fillings
  4. although if you don’t rub it in it’ll stain the mankini
  5. clearly it was meant to say “ahahaha responsive design, what evaaaaar”, but maybe that didn’t fit
Author: "sil"
Send by mail Print  Save  Delicious 
Date: Wednesday, 17 Sep 2014 13:07

Windows applications sometimes fail to load. But why? It’ll not tell you, it will instead show a generic and pointless “Application Error” message. Inside this message you will read something like this:

The application was unable to start correctly (0xc0000142). Click OK to close the application.

The only thing you can do here is close the application and search on the Internet for that cryptic error code. And maybe it’s the reason why you are reading this post.
It’s not that easy to find a solution to this problem, but I found it thanks to Up and Ready and want to share it with you.

The problem

Windows tells you that the application was unable to start. You can try a hundred times, but the error does not solve itself magically, because it’s not casual. The problem is that the ddl that launches the application is unsigned or digitally no longer valid. And it’s not up to you, maybe you just downloaded the program from the official site.

The solution

To solve the Application Error you need an advanced Windows Sysinternals Tool called Autoruns for Windows. You can download it from the official website.

Windows Application Error Autoruns AppInit

Click on the image to view it full size.

Extract the archive you downloaded, launch autoruns.exe and go to the AppInit tab, which will list all the dll that are unsigned or digitally no longer valid on you computer. Right click each of them, one at a time, go to Properties and rename them. After renaming each of them, try launching the application again to find the problematic dll.

If the previous method didn’t solve the application error, right click on the following entry:

HKLM\SOFTWARE\Microsoft\Windows NT\CurrentVersion\Windows\AppInit_Dlls

and click on Jump to entry…

Windows Application Error System Registry Editor

A new window opens: it’s the System Registry Editor. Double click LoadAppInit_DLLs and change the value from 1 to 0. Click OK to confirm and exit. Now launch the compromised program and it’ll start.

Note: some applications may change that value back to 1 after they get launched!

The post Windows: How to Solve Application Error 0xc0000142 and 0xc0000005 appeared first on deshack.

Author: "Mattia “deshack” Migliorini"
Send by mail Print  Save  Delicious 
Date: Wednesday, 17 Sep 2014 05:03

Linux creator Linus Torvalds is well-known for his strong opinions on many technical things. But when it comes to systemd, the init system that has caused a fair degree of angst in the Linux world, Torvalds is neutral.

“When it comes to systemd, you may expect me to have lots of colourful opinions, and I just don’t,” Torvalds told iTWire in an interview. “I don’t personally mind systemd, and in fact my main desktop and laptop both run it.

Source:

http://www.itwire.com/business-it-news/open-source/65402-torvalds-says-he-has-no-strong-opinions-on-systemd

Submitted by: Sam Varghese

Author: "kr1st0"
Send by mail Print  Save  Delicious 
Date: Tuesday, 16 Sep 2014 17:15

Meeting Minutes

IRC Log of the meeting.

Meeting minutes.

Agenda

20140916 Meeting Agenda


Release Metrics and Incoming Bugs

Release metrics and incoming bug data can be reviewed at the following link:

  • http://people.canonical.com/~kernel/reports/kt-meeting.txt


Status: Utopic Development Kernel

The Utopic kernel remains based on a v3.16.2 upstream stable kernel and
is uploaded to the archive, ie. linux-3.16.0-15.21. Please test and let
us know your results.
I’d also like to point out that our Utopic kernel freeze date is about 3
weeks away on Thurs Oct 9. Please don’t wait until the last minute to
submit patches needing to ship in the Utopic 14.10 release.
—–
Important upcoming dates:
Mon Sep 22 – Utopic Final Beta Freeze (~1 weeks away)
Thurs Sep 25 – Utopic Final Beta (~1 weeks away)
Thurs Oct 9 – Utopic Kernel Freeze (~3 weeks away)
Thurs Oct 16 – Utopic Final Freeze (~4 weeks away)
Thurs Oct 23 – Utopic 14.10 Release (~5 weeks away)


Status: CVE’s

The current CVE status can be reviewed at the following link:

http://people.canonical.com/~kernel/cve/pkg/ALL-linux.html


Status: Stable, Security, and Bugfix Kernel Updates – Trusty/Precise/Lucid

Status for the main kernels, until today (Sept. 16):

  • Lucid – verification & testing
  • Precise – verification & testing
  • Trusty – verification & testing

    Current opened tracking bugs details:

  • http://kernel.ubuntu.com/sru/kernel-sru-workflow.html

    For SRUs, SRU report is a good source of information:

  • http://kernel.ubuntu.com/sru/sru-report.html

    Schedule:

    cycle: 29-Aug through 20-Sep
    ====================================================================
    29-Aug Last day for kernel commits for this cycle
    31-Aug – 06-Sep Kernel prep week.
    07-Sep – 13-Sep Bug verification & Regression testing.
    14-Sep – 20-Sep Regression testing & Release to -updates.


Open Discussion or Questions? Raise your hand to be recognized

No open discussion.

Author: "Joseph Salisbury"
Send by mail Print  Save  Delicious 
Date: Tuesday, 16 Sep 2014 17:01

Last week I flew out to the east coast to attend the very first Fossetcon. The conference was on the smaller side, but I had a wonderful time meeting up with some old friends, meeting some new Ubuntu enthusiasts and finally meeting some folks I’ve only communicated with online. The room layout took some getting used to, but the conference staff was quick to put up signs and directing conference attendees in the right direction and in general leading to a pretty smooth conference experience.

On Thursday the conference hosted a “day zero” that had training and an Ubucon. I attended the Ubucon all day, which kicked off with Michael Hall doing an introduction to the Ubuntu on Phones ecosystem, including Mir, Unity8 and the Telephony features that needed to be added to support phones (voice calling, SMS/MMs, Cell data, SIM card management). He also talked about the improved developer portal with more resources aimed at app developers, including the Ubuntu SDK and simplified packaging with click packages.

He also addressed the concern of many about whether Ubuntu could break into the smartphone market at this point, arguing that it’s a rapidly developing and changing market, with every current market leader only having been there for a handful of years, and that new ideas need need to play to win. Canonical feels that convergence between phone and desktop/laptop gives Ubuntu a unique selling point and that users will like it because of intuitive design with lots of swiping and scrolling actions, gives apps the most screen space possible. It was interesting to hear that partners/OEMs can offer operator differentiation as a layer without fragmenting the actual operating system (something that Android struggles with), leaving the core operating system independently maintained.

This was followed up by a more hands on session on Creating your first Ubuntu SDK Application. Attendees downloaded the Ubuntu SDK and Michael walked through the creation of a demo app, using the App Dev School Workshop: Write your first app document.

After lunch, Nicholas Skaggs and I gave a presentation on 10 ways to get involved with Ubuntu today. I had given a “5 ways” talk earlier this year at the SCaLE in Los Angeles, so it was fun to do a longer one with a co-speaker and have his five items added in, along with some other general tips for getting involved with the community. I really love giving this talk, the feedback from attendees throughout the rest of the conference was overwhelmingly positive, and I hope to get some follow-up emails from some new contributors looking to get started. Slides from our presentation are available as pdf here: contributingtoubuntu-fossetcon-2014.pdf


Ubuntu panel, thanks to Chris Crisafulli for the photo

The day wrapped up with an Ubuntu Q&A Panel, which had Michael Hall and Nicholas Skaggs from the Community team at Canonical, Aaron Honeycutt of Kubuntu and myself. Our quartet fielded questions from moderator Alexis Santos of Binpress and the audience, on everything from the Ubuntu phone to challenges of working with such a large community. I ended up drawing from my experience with the Xubuntu community a lot in the panel, especially as we drilled down into discussing how much success we’ve had coordinating the work of the flavors with the rest of Ubuntu.

The next couple days brought Fossetcon proper, with I’ll write about later. The Ubuntu fun continued though! I was able to give away 4 copies of The Official Ubuntu Book, 8th Edition which I signed, and got José Antonio Rey to sign as well since he had joined us for the conference from Peru.

José ended up doing a talk on Automating your service with Juju during the conference, and Michael Hall had the opportunity to a talk on Convergence and the Future of App Development on Ubuntu. The Ubuntu booth also looked great and was one of the most popular of the conference.

I really had a blast talking to Ubuntu community members from Florida, they’re a great and passionate crowd.

Author: "pleia2"
Send by mail Print  Save  Delicious 
Date: Tuesday, 16 Sep 2014 15:24

Hi, after a lot of work, thinking and talking about the problem of the LoCo Organization and the SubLoCos, we came up with the following policy:

  • Each team will be a country (or state in the United States). We will call this a ‘LoCo’.
  • Each LoCo can have sub-teams. This sub-teams will be created at the will and need of each LoCo.
  • A LoCo may have sub-teams or not have sub-teams.
  • In the event a LoCo does have sub-teams, a Team Council needs to be created.
  • A Team Council is conformed by at least one member of each sub-team.
  • The members that will be part of the Team Council will be chosen by other current members of the team.
  • The Team Council will have the power to make decisions regarding to the LoCo.
  •  The Team Council will also have the power to request partner items, such as conference and DVD packs.
  • The LoCo Council will only recognize one team per country (or state in the United States). This is the team that will be in the ~locoteams team in Launchpad.
  • In the event a LoCo wants to go through the verification process, the LoCo will go through it, and not individual sub-teams.
  • LoCos not meeting the criteria of country/state teams will be denied verification.
  • In the event what is considered a sub-team wants to be considered a LoCo, it will need to present a request to the LoCo Council.
  • The LoCo Council will provide a response, which is, in no way, related to verification. The LoCo will still have to apply for verification if wanted.

We encourage the LoCo teams to see if this new form of organization is fits for you, if so please start forming subteams as you find useful. If a team needs help with this or anything else contact us, we are here to help!

Author: "Pablo Rubianes"
Send by mail Print  Save  Delicious 
Date: Monday, 15 Sep 2014 23:51

Welcome to the Ubuntu Weekly Newsletter. This is issue #383 for the week September 8 – 14, 2014, and the full version is available here.

In this issue we cover:

The issue of The Ubuntu Weekly Newsletter is brought to you by:

  • Elizabeth K. Joseph
  • Jose Antonio Rey
  • And many others

If you have a story idea for the Weekly Newsletter, join the Ubuntu News Team mailing list and submit it. Ideas can also be added to the wiki!

Except where otherwise noted, content in this issue is licensed under a Creative Commons Attribution 3.0 License BY SA Creative Commons License

Author: "lyz"
Send by mail Print  Save  Delicious 
Date: Monday, 15 Sep 2014 16:22

On a 15th September, 3 years ago, I got my Ubuntu Membership.

There’s only thing I can say about it: it’s been the most wonderful and awesome 3 years I could have. I would’ve never thought that I would find such welcoming and amazing community.

Even though I may have not worked with you directly, thank you. You all are what makes the community awesome – I wouldn’t imagine it without one of you. We are all building the future, so let’s continue!

As I said on the title, I hope that it’s not only 3 years. I’ll keep on counting!


Author: "José Antonio Rey"
Send by mail Print  Save  Delicious 
Date: Monday, 15 Sep 2014 14:50

Back in April, I upstreamed (that is, reported a bug to Debian) regarding the `nginx-naxsi` packages. The initial bug I upstreamed was about the outdated naxsi version in the naxsi packages. (see this bug in Ubuntu and the related bug in Debian)

The last update on the Debian bug is on September 10, 2014. That update says the following, and was made by Christos Trochalakis:

After discussing it with the fellow maintainers we have decided that it is
better to remove the nginx-naxsi package before jessie is freezed.

Packaging naxsi is not trivial and, unfortunately, none of the maintainers uses
it. That’s the reason nginx-naxsi is not in a good shape and we are not feeling
comfortable to release and support it.

We are sorry for any inconvenience caused.

I asked what the expected timeline was for the packages being dropped. In a response from Christos today, September 15, 2014, it was said:

It ‘ll get merged and released (1.6.1-3) by the end of the month.


In Ubuntu, these changes will likely not make it into 14.10, but future versions of Ubuntu beyond 14.10 (such as 15.04) will likely have this change.

In the PPAs, the naxsi packages will be dropped with stable 1.6.1-3+precise0 +trusty0 +utopic0 and mainline 1.7.4-1+precise0 +trusty0 +utopic0 or will be dropped in later versions if a new point release is made before then.

In Debian, these changes are likely to hit by the end of the month (with 1.6.1-3).

Author: "teward"
Send by mail Print  Save  Delicious 
Date: Monday, 15 Sep 2014 09:00

Last week I attended FOSSETCON, a new open source convention here in central Florida, and I had the opportunity to give a couple of presentations on Ubuntu phones and app development. Anybody who knows me knows that I love talking about these things, but a lot fewer people know that doing it in front of a room of people I don’t know still makes me extremely nervous. I’m an introvert, and even though I have a public-facing job and work with the wider community all the time, I’m still an introvert.

I know there are a lot of other introverts out there who might find the idea of giving presentations to be overwhelming, but they don’t have to be.  Here I’m going to give my personal experiences and advice, in the hope that it’ll encourage some of you to step out of your comfort zones and share your knowledge and talent with the rest of us at meetups and conferences.

You will be bad at it…

Public speaking is like learning how to ride a bicycle, everybody falls their first time. Everybody falls a second time, and a third. You will fidget and stutter, you will lose your train of thought, your voice will sound funny. It’s not just you, everybody starts off being bad at it. Don’t let that stop you though, accept that you’ll have bruises and scrapes and keep getting back on that bike. Coincidentally, accepting that you’re going to be bad at the first ones makes it much less frightening going into them.

… until you are good at it

I read a lot of things about how to be a good and confident public speaker, the advice was all over the map, and a lot of it felt like pure BS.  I think a lot of people try different things and when they finally feel confident in speaking, they attribute whatever their latest thing was with giving them that confidence. In reality, you just get more confident the more you do it.  You’ll be better the second time than the first, and better the third time than the second. So keep at it, you’ll keep getting better. No matter how good or bad you are now, you will keep getting better if you just keep doing it.

Don’t worry about your hands

You’ll find a lot of suggestions about how to use your hands (or not use them), how to walk around (or not walk around) or other suggestions about what to do with yourself while you’re giving your presentation. Ignore them all. It’s not that these things don’t affect your presentation, I’ll admit that they do, it’s that they don’t affect anything after your presentation. Think back about all of the presentations you’ve seen in your life, how much do you remember about how the presenter walked or waved their hands? Unless those movements were integral to the subject, you probably don’t remember much. The same will happen for you, nobody is going to remember whether you walked around or not, they’re going to remember the information you gave them.

It’s not about you

This is the one piece of advice I read that actually has helped me. The reason nobody remembers what you did with your hands is because they’re not there to watch you, they’re there for the information you’re giving them. Unless you’re an actual celebrity, people are there to get information for their own benefit, you’re just the medium which provides it to them.  So don’t make it about you (again, unless you’re an actual celebrity), focus on the topic and information you’re giving out and what it can do for the audience. If you do that, they’ll be thinking about what they’re going to do with it, not what you’re doing with your hands or how many times you’ve said “um”. Good information is a good distraction from the things you don’t want them paying attention to.

It’s all just practice

Practicing your presentation isn’t nearly as stressful as giving it, because you’re not worried about messing up. If you mess up during practice you just correct it, make a note to not make the same mistake next time, and carry on. Well if you plan on doing more public speaking there will always be a next time, which means this time is your practice for that one. Keep your eye on the presentation after this one, if you mess up now you can correct it for the next one.

 

All of the above are really just different ways of saying the same thing: just keep doing it and worry about the content not you. You will get better, your content will get better, and other people will benefit from it, for which they will be appreciative and will gladly overlook any faults in the presentation. I guarantee that you will not be more nervous about it than I was when I started.

Author: "Michael Hall"
Send by mail Print  Save  Delicious 
Date: Monday, 15 Sep 2014 08:23

Last week’s autopkgtest 3.5 release (in Debian sid and Ubuntu Utopic) brings several new features which I’d like to announce.

Tests that reboot

For testing low-level packages like init or the kernel it is sometimes desirable to reboot the testbed in the middle of a test. For example, I added a new boot_and_services systemd autopkgtest which configures grub to boot with systemd as pid 1, reboots, and then checks that the most important services like lightdm, D-BUS, NetworkManager, and cron come up as expected. (This test will be expanded a lot in the future to cover other areas like the journal, logind, etc.)

In a testbed which supports rebooting (currently only QEMU) your test will now find an “autopkgtest-reboot” command which the test calls with an arbitrary “marker” string. autopkgtest will then reboot the testbed, save/restore any files it needs to (like the tests file tree or previously created artifacts), and then re-run the test with ADT_REBOOT_MARK=mymarker.

The new “Reboot during a test” section in README.package-tests explains this in detail with an example.

Implicit test metadata for similar packages

The Debian pkg-perl team recently discussed how to add package tests to the ~ 3.000 Perl packages. For most of these the test metadata looks pretty much the same, so they created a new pkg-perl-autopkgtest package which centralizes the logic. autopkgtest 3.5 now supports an implicit debian/tests/control control file to avoid having to modify several thousand packages with exactly the same file.

An initial run already looked quite promising, 65% of the packages pass their tests. There will be a few iterations to identify common failures and fix those in pkg-perl-autopkgtest and autopkgtestitself now.

There is still some discussion about how implicit test control files go together with the DEP-8 specification, as other runners like sadt do not support them yet. Most probably we’ll declare those packages XS-Testsuite: autopkgtest-pkg-perl instead of the usual autopkgtest.

In the same vein, Debian’s Ruby maintainer (Antonio Terceiro) added implicit test control support for Ruby packages. We haven’t done a mass test run with those yet, but their structure will probably look very similar.

Author: "pitti"
Send by mail Print  Save  Delicious 
Date: Monday, 15 Sep 2014 07:00

Hi all,
after long time I return to write to show you how to create a simple game for Ubuntu for Phones (but also for Android) with Bacon2D.

Bacon2D is a framework to ease 2D game development, providing ready-to-use QML elements representing basic game entities needed by most of games.

As tutorial I’ll explain you how I create my first QML game, 100balls, that you could find on Ubuntu Store on Phones. Source is available on Github.

Installation

So, first of all we need to install Bacon2D on our system. I suppose you have already installed QT on your system, so we only need to take source and compile it:

git clone git@github.com:Bacon2D/Bacon2D.git
cd Bacon2D
mkdir build && cd build
qmake ..
make
sudo make install

Now you have Bacon2D on your system, and you can import it in every project you want.

A first look to Bacon2D

Bacon2D provides a good number of custom components for your app. Of course, I can’t describe them all in one article, so please read the documentation. We’ll use only few of them, and I think the best way to introduce you to them is writing the app.
So, let’s start!

First of all, we create our base file, called 100balls.qml:

import QtQuick 2.0
import Bacon2D 1.0

The first element we add is the Game element. Game is the top-level container, where all the game will be. We set some basic property and the name of the game, with gameName property:

import QtQuick 2.0
import Bacon2D 1.0
 
Game {
    id: game
    anchors.centerIn: parent
 
    height: 680
    width: 440
 
    gameName: "com.ubuntu.developer.rpadovani.100balls" // Ubuntu Touch name format, you can use whatever you want
}

But the Game itself is useless, we need to add one or more Scene to it. A scene is the place where all Entity of the game will be placed.
Scene has a lot of property, for now is importat to set two of them: running indicates if all things in the scene will move, and if game engine works; second property is physics, that indicates if Box2D has to be used to simulate physic in the game. We want a game where some balls fall, so we need to set it to true.

import QtQuick 2.0
import Bacon2D 1.0
 
Game {
    id: game
    anchors.centerIn: parent
 
    height: 680
    width: 440
 
    gameName: "com.ubuntu.developer.rpadovani.100balls" // Ubuntu Touch name format, you can use whatever you want
 
    Scene {
        id: gameScene
        physics: true
        running: true
    }
}
Author: "Riccardo Padovani"
Send by mail Print  Save  Delicious 
Date: Sunday, 14 Sep 2014 23:56

Today I released the first of the Brum Tech Scene interviews, with me talking to Simon Jenner of Silicon Canal and Oxygen Startups. There’s a video on the site from me explaining why I’m doing this, but I figure that the more discerning audience for as days pass by might appreciate a more in-depth discussion.

I love this city. I love that we’re prepared to spend a hundred and ninety million quid on building the best library in the whole world. I love that there’s so much going on, tech-wise. But nobody talks to anybody else. If you look at, say, Brighton, the whole tech scene there all hang out together. They can put on a Digital Brighton week and have dConstruct be part of it and Seb do mad things with visualisations and that’s marvellous. We ought to have that. I want us to have that.

We don’t have a tech scene. We’ve got twenty separate tech scenes. What I want to do is knock down the walls a bit. So the designers talk to the SEO people and the Linux geeks talk to the designers. Because there is no way that this can be a bad thing.

I also want to learn a bit about videos. Now, let’s be clear here. I know from a decade of podcasting that with a mild expenditure of money on gear, and a great sound engineer (Jono Bacon, step forward) you can produce something as good as the professionals. Bad Voltage sounds as good, production-wise, as the BBC’s Today programme does. Video is not like that. There is a substantial difference between amateur and professional efforts; one bloke using mobile phones to record cannot make something that looks like Sherlock or Game of Thrones. I’m not trying to look professional here; I’m aiming for “competent amateur”. I’ve learned loads about how to record a video interview, how to mix it, how to do the editing. Sit far enough apart that your voice doesn’t sound on their mic. Apply video effects to the clip before you cut it up. Don’t speak over the interviewee. KDEnLive’s “set audio reference” is witchcraft brilliance. I knew none of this two months ago. And I’ve really enjoyed learning. I am in no wise good at this stuff, but I’m better than I was.

This has been a fun project to set up, and it will continue being fun as I record more interviews. My plan is to have a new one every Monday morning, indefinitely, as long as people like them and I’m still interested in doing them. I should give big love to Mike, my designer, who I fought with tooth and nail about the site design and the desaturated blue look to the videos, and to Dan Newns who sat and was interviewed as a test when I first came up with this idea, and has provided invaluable feedback throughout.

If you know something about video editing, I’d love to hear how I can do better. Ping me on twitter or by mail. Tell me as well who you want to hear interviewed; which cool projects are going on that I don’t know about. I’d also love to hear about cool venues in the city in which I can do interviews; one of my subsidiary goals here is to show off the city’s tech places. Annoyingly, I spoke to the Library and to the Birmingham Museums Trust and they were all “fill out our fifteen page form” because they’re oriented around the BBC coming in with a crew of twenty camera people, not one ginger guy with a mobile phone and a dream. Maybe I’ll do things with @HubBirmingham once they actually exist.

I should talk about the tech, here. I record the interviews on an iPhone 5, a Nexus 4, and a little HD camera I bought years ago. The audio is done with two Røde Smartlav lapel mics plugged into the two phones. None of this is expensive, which has a cost in terms of video and audio quality but critically doesn’t have much of a cost in terms of actual pounds sterling. And editing is done with KDEnLive (kdenlive?) which is a really powerful non-linear video editor for Ubuntu, and the team who make it should be quite proud. The big thing I’m missing (apart from a cameraman) is a tripod, which I can probably buy for about ten quid, and I will do once I find one that’s tall and yet still fits in my laptop bag.

Anyway, that’s the story of the Brum Tech Scene interviews. There’ll be one every Monday. I hope you like them. I hope they help, even in a small way, to make the Brum tech scene gel together even more than it has thus far. Let me know what you think. brumtechscene.co.uk.

Author: "sil"
Send by mail Print  Save  Delicious 
Date: Sunday, 14 Sep 2014 23:24

I will start this off by saying: I’m very (and honestly) sorry for, well, everything.

To give a bit of history, I started relinux as a side-project for my CosmOS project (cloud-based distribution … which failed), in order to build the ISO’s. The only reasonable alternative at the time was remastersys, and I realized I would have to patch it anyways, so I thought that I might as well make a reusable tool for other distributions to use too.

Then came a rather large amount of friction between me and the author of remastersys, of which I will not go into any detail of. I acted very immaturely then, and wronged him several times. I had defamed him, made quite a few people very angry at him, and even managed to get some of his supporters against him. True, age and maturity had something to do with it (I was 12 at the time), but that still doesn’t excuse my actions at all.

So my first apology is to Tony Brijeski, the author of remastersys, for all the trouble and possible pain I had put him through. I’m truly sorry for all of this.

However, though the dynamics with Tony and remastersys are definitely a large part of why I’m quitting relinux, that is not all. The main reason, actually, is lack of interest. I have rewritten relinux a total of 7 times (including the original fork of remastersys), and I really hate the debugging process (takes 15-20 minutes to create an ISO, so that I can debug it). I have also lost interest in creating linux distributions, so not only am I very tired of working on it, I also don’t really care about what it does.

On this note, my second apologies (and thanks) have to go those who have helped me so much through the process, especially those who have tried to encourage me to finish relinux. Those listed are in no particular order, and if I forgot you, then let me know (and I apologize for that!):

  • Ko Ko Ye
  • Raja Genupula
  • Navdeep Sidhu
  • Members of the TSS Web Dev Club
  • Ali Hallahi
  • Gert van Spijker
  • Aritra Das
  • Diptarka Das
  • Alejandro Fernandez
  • Kendall Weaver

Thank you very much for everything you’ve done!

Lastly, I would like to explain my plans for it, in case anyone wants to continue it (by no means do I want to enforce these, these are just ideas).

My plan for the next release of relinux was to actually make a very generic and scriptable CLI ISO creation tool, and then make relinux as a specific set of “profiles” for that tool (plus an interface). The tool would basically contain a few libraries for the chosen scripting language, for things like storing the filesystem (SquashFS or other), ISO creation, and general utilities for editing files while keeping permissions, mutli-threading/processing, etc… The “profiles” would then copy, edit, and delete files as needed, set up the tool wanted for running the live system (in ubuntu’s case, this’d be casper), setup the installer/bootloader, and such.

I would like to apologize to you all, the people who have used relinux and have waited for a stable version for 3 years, for not doing this. Thank you very much for your support, and I’m very sorry for having constantly pushed releases back and having never made a stable or well working version of relinux. Though I do have some excuses as to why the releases didn’t work, or why I didn’t test them well enough, none of them can cover why I didn’t fix them or work on it more. And for that, I am very sorry.

I know that this is a very large post for something so simple, but I feel that it would not be right if I didn’t apologize to those I have done wrong to, and thanked those who have helped me along the way.

So to summarize, thank you, sorry, and relinux is now dead.

- Joel Leclerc (MiJyn)


Author: "MiJyn"
Send by mail Print  Save  Delicious 
Date: Sunday, 14 Sep 2014 20:07

My last post was about getting started in a career in information security. This post is about the sport end of information security: Capture the Flag (CTFs).

I'd played around with some wargames (Smash the Stack, Over the Wire, and Hack this Site) before, but my first real CTF (timed, competitive, etc.) was the CTF run by Mad Security at BSides SF 2013. By some bizarre twist of fate, I ended up winning the CTF, and I was hooked. I've probably played in about 30 CTFs since, most of them online with the team Shadow Cats. It's been a bumpy ride, but I've learned a lot about a variety of topics by doing this.

If you're in the security industry and you've never tried a CTF, you really should. Personally, I love CTFs because they get me to exercise skills that I never get to use at work. They also inspire some of my research and learning. The only problem is making the time. :)

Here's some resources I've thought were interesting:

Author: "David Tomaschik"
Send by mail Print  Save  Delicious 
Date: Saturday, 13 Sep 2014 20:43
I look forward to Ubuntu's semiannual release day, because it's the completion of 6ish months of work by Ubuntu (and by extension Debian) developers.

I also loathe it, because every single time we get people saying "This Ubuntu release is the worst release ever!".

Ubuntu releases are always rocky around release time, because the first time Ubuntu gets widespread testing is on or after release day.

We ship software to 12 Million Ubuntu Users with only 150 MOTUs who work directly on the platform. That's a little less than 1 developer with upload rights to the archive for every 60,000 users. ((This number, like all other usage data, is dated, and probably wasn't even accurate when it was first calculated)) Compared to Debian, which (at last estimate in 2010) had 1.5 million uniques on security.debian.org, yet has around 1000 Debian Developers.

Debian has a strong testing culture; someone once estimated that around ¾ of Debian users are running unstable or testing. In Ubuntu, we don't have good metrics on how many people are using the development release that I'm aware of (pointers welcome), but I'd guess that it's a very very small percentage. A common thread in bug reports, if we get a response at all, goes on as follows:
Triager: ((Developer, bugcontrol member, etc. Somebody who is not experiencing the problem but wants to help.)) "Is this a problem in $devel?"
User: "I'll let you know when it hits final"
Triager: "It's too late then. Then we'll want you to test in the next release. We have to fix it BEFORE its final"
User: "Ok, I'll test at beta."
Triager: "That's 2 weeks before release, which will be too late. Please test ASAP if you want us to have time to fix it"

Of course, there are really important bugs with hardware support which keep on cropping up. But if they're just getting reported on or around release day, there are limits to what can be done about them this cycle.

We need to make it easier for people to run early development versions, and encourage more people to use them (as long as they're willing to deal with breakage). I'm not sure whether unstable/testing is appropriate for Ubuntu, and I'm fairly confident that we don't want to move to a rolling release (currently being discussed in Debian, summary). But we badly need more developers, and equally importantly, more testers to try it out earlier in the release process.

To users: please, please try out the development versions. Download a LiveCD and run a smoketest, or check if bugs you reported are in fact fixed in the later versions. And do it early and often.
Author: "Luke Faraone"
Send by mail Print  Save  Delicious 
Date: Saturday, 13 Sep 2014 19:30

I've only been an information security practitioner for about a year now, but I've been doing things on my own for years before that. However, many people are just getting into security, and I've recently stumbled on a number of resources for newcomers, so I thought I'd put together a short list.

Author: "David Tomaschik"
Send by mail Print  Save  Delicious 
Date: Saturday, 13 Sep 2014 15:51

When you talk about the “user experience” of the thing you’re building, remember that developers who use your APIs are users too. And you need to think about their experience.

We seem to have created a world centred on github where everyone has to manage dependencies by hand, like we had to in 1997. This problem was completely solved by apt twenty years ago, but the new cool github world is, it seems, too cool to care about that. Go off to get some new project by git cloneing it and it’s quite likely to say “oh, and it depends on $SOME_OTHER_PROJECT (here’s a link to that project’s github repo)”. And then you have to go fetch both and set them up yourself. Which is really annoying.

Now, there are good reasons why to not care about existing dependency package management systems such as apt. Getting stuff into Ubuntu is hard, laborious work and most projects don’t want to do it. PPAs make it easier, but not much easier; if you’re building a thing and not specifically targeting Ubuntu with it, you don’t want to have to learn about Launchpad and PPAs and build recipes and whatnot. This sort of problem is also solves neatly for packages in a specific language by that language’s own packaging system; Python stuff is installable with pip install whatever and a virtualenv; Node stuff is installable with npm install whatever; all these take care of fetching any dependent stuff. But this rush for each language to have its own “app store” for its apps and libraries means that combining things from different languages is still the same 20th century nightmare. Take, for example, Mozilla’s new Firefox Tools Adaptor. I’m not picking on Mozilla here; the FTA is new, and it’s pretty cool, and it’s not finished yet. This is just the latest in a long line of things which exhibit the problem. The FTA allows you to use the Firefox devtools to debug web things running in other browsers. Including, excitingly, debugging things running in iOS Safari on the iPhone. Now, doing that’s a pain in the ringpiece at the moment; you have to install Google’s ios-webkit-debug-proxy, which needs to be compiled, and Apple break compatibility with it all the time and so you have to fetch and build new versions of libimobiledevice or something. I was eager to see that the new Firefox Tools Adaptor promises to allow debugging on iOS Safari just by installing a Firefox extension.

And then I read about it, and it says, “The Adapter’s iOS support uses Google’s ios-webkit-debug-proxy. Until that support is built directly into the add-on, you’ll need to install and run the ios-webkit-debug-proxy binary yourself”. Sigh. That’s the hard part. And it’s not any easier here.

Again, I’m not blaming Mozilla here — they plan to fix this, but they’ll have to fix it by essentially bundling ios-webkit-debug-proxy with the FTA. That’ll work, and that’s an important thing for them to do in order to provide a slick user experience for developers using this tool (because “download and compile this other thing first” is not ever ever a nice user experience).

It is made worse by people using a language packaging system (designed for people developing libraries for a given language) to do app distribution. See, for example, tmuxme, which is an app for sharing a terminal session with many people (think of it like screen sharing, but for a terminal). And how do you install it? gem install tmuxme. No. Ruby’s gem command is for developers to download a Ruby library that their Ruby package needs. I, as someone who wants to use this tool, should not have to care that it’s written in Ruby. I should not have to have a Ruby development environment set up in order to use an app. See the birmingham.io forum thread for much much more about this, and why it doesn’t even work. New rather cool app pup is the same — it’s a little app, inspired by the excellent jq, into which I can pipe HTML and give it a CSS selector, and pup will then print just the elements which match the selector. But how do I install it? go get github.com/ericchiang/pup. No. I don’t have go. I don’t have a go environment set up. I don’t have $GOPATH set. I shouldn’t even have to care that this little util is even written in Go. It’s a utility. What’s worse about this is that, unlike Ruby or Python, Go creates actual executables; I don’t even need the Go system around to run it! Why should I need to install all of Go just to get your app? Don’t use a language-specific library packaging system for distribution of applications. Don’t make me identify and download dependencies myself just because you already have them.

This is sorta kinda solved by brew for Mac users, but there’s a lot of stuff not in brew either. Still, there is willingness to solve it that way by having a packaging system. But it’s annoying that Ubuntu already has one and people are loath to use it. Using it makes for a better developer user experience. That’s important.

Author: "sil"
Send by mail Print  Save  Delicious 
Next page
» You can also retrieve older items : Read
» © All content and copyrights belong to their respective authors.«
» © FeedShow - Online RSS Feeds Reader