By default the Qt5 libraries distributed from qt.io are dynamically linked. This means that every Qt app dynamically references and uses the Qt prebuild libraries (.dll or .so) it depends on. So in order to deploy your Qt app to your users, you need to find the relevant Qt libraries and distribute them as well (or make sure all other PCs have exactly the same Qt environment as your development PC), which is a pain. That is why it’s easier to build a static version of your application – one single standalone executable with all libraries included inside. But to build a standalone executable of an app you need a static version of Qt libraries. If you target Windows 7/8, you might want to use the Powershell described in qt.io documentation to build a static Qt for Windows, but you can also compile a static Qt5 from scratch on your own. And no, it’s not that difficult…
Category: Open Source Page 2 of 10
Another release has just emerged from the depths of the Social Network Visualizer project! The new version 1.9 fixes important bugs and brings a vastly improved matrix inverse algorithm, thanks to the newly implemented LU decomposition. Of course, the new algorithm speeds up all other routines which depend on the inverse of a matrix, such as the Information Centrality algorithm.
Also, SocNetV v1.9 brings a revamped -and corrected- PageRank Prestige routine. Up to v1.8, the PageRank algorithm was using the original Page&Brin formula which lead to somewhat different results compared to other SNA applications. From this version, SocNetV uses the correct formula and computes comparable results (yes, we tested :)).
Source code, Windows zipped executables, Mac OS image and binary packages for major Linux distributions are as always available from the project’s Downloads page. Have fun with the new release!
Also, I am happy to report that the Qt project has added SocNetV to its ‘showroom’ page, along with many other nice Qt applications. There is also a showroom page dedicated to SocNetV with nice screenshots and a brief description of the application. Kudos to Kojo Tero, of Qt Company, for this!
One of the daily routines administrators do is to monitor their web server logs for various interesting things: response codes (i.e. 500), attempts to access restricted pages, user-agents, ips, popular pages, even image hotlinking from other sites. In Linux servers, this can be easily done through shell one-liners involving various tools (tail, awk, sort, cut etc). Here is a bash script I use which automates apache monitoring for a given website.
If you use Git for version controlling some projects of yours and you are a number-crunch fanboy, you might be interested to know which hours you (or some fellows in your team) are active the most. I mean which hours during the day you do your commits. This is onliner will do this for you. Run it from inside your project dir: