Skip to main content

Eclipse TPTP on IBM AIX platform

It seems that the Eclipse TPTP project has stopped to provide Agent Controller and agent libraries for some Unix platforms like IBM AIX.

I have just tried to use the IBM Rational Agent Controller (available at http://www-01.ibm.com/support/docview.wss?rs=2042&uid=swg27013420) with Eclipse TPTP 4.7.1 as client. Surprisingly it seems to be perfectly compatible, i have successfully collected profiling data on an Eclipse Workbench under Windows from a IBM J9 32bits VM SR8 running on a 64bits IBM AIX PowerPC 6.

Some tips :

  • While i tried to profile with an Xmx value equals or greater than 2048m the JVM crashed  with the following error as soon as i attached the Eclipse Workbench to the Agent Controller:
Unhandled exception
Type=Illegal instruction vmState=0x00000000
J9Generic_Signal_Number=00000010 Signal_Number=00000004 Error_Value=00000000 Signal_Code=0000001e
…..
Target=2_40_20100401_055940 (AIX 5.3)
CPU=ppc (4 logical CPUs) (0x200000000 RAM)
----------- Stack Backtrace -----------
---------------------------------------
JVMDUMP006I Processing dump event "gpf", detail "" - please wait.
JVMDUMP032I JVM requested System dump using 'core.20101013.171437.2584664.0001.dmp' in response to an event
Note: "Enable full CORE dump" in smit is set to FALSE and as a result there will be limited threading information in core file.
JVMDUMP006I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" - please wait.
JVMDUMP006I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" - please wait.
JVMDUMP010I System dump written to core.20101013.171437.2584664.0001.dmp
JVMDUMP032I JVM requested Snap dump using 'Snap.20101013.171437.2584664.0004.trc' in response to an event
JVMDUMP010I Snap dump written to Snap.20101013.171437.2584664.0004.trc
JVMDUMP032I JVM requested Java dump using 'javacore.20101013.171437.2584664.0005.txt' in response to an event
JVMDUMP032I JVM requested Snap dump using 'Snap.20101013.171440.2584664.0003.trc' in response to an event
JVMDUMP010I Snap dump written to Snap.20101013.171440.2584664.0003.trc
JVMDUMP032I JVM requested Snap dump using Snap.20101013.171440.2584664.0002.trc' in response to an event
JVMDUMP010I Snap dump written to Snap.20101013.171440.2584664.0002.trc
UTE430: can't allocate buffer
UTE437: Unable to load formatStrings for j9mm
JVMDUMP032I JVM requested Heap dump using 'heapdump.20101013.171440.2584664.0006.phd' in response to an event
JVMDUMP010I Java dump written to javacore.20101013.171437.2584664.0005.txt
JVMDUMP013I Processed dump event "gpf", detail "".

  • Do not forget to properly setup the LIBPATH environment variable in order to reference the lib and plugins/org.eclipse.tptp.javaprofiler directories. Agents rely on it to find the *.so files.
  •  Do not forget to start the AgentController on the AIX (with the ACStart.sh script in the bin directory) and, if it is not already done, to configure it with the SetConfig.sh Before trying to profile i suggest you to test the Agent Controller from a remote client with the SampleClient :
SampleClient <AIXHostName> <AgentControllerPort>

(Default AgentControllerPort is 10006)

Comments

Popular posts from this blog

Orientée colonnes ?

Les bases NoSQL sont arrivées avec leur cortège de nouveautés et pour certaines d'entre elles une notion héritée de BigTable : celle de base de donnée orientée colonne. Cependant faire le lien entre l'article de Wikipedia et comprendre ce que permet réellement un base de donnée comme HBase n'est pas une chose évidente. En effet le simple fait de définir cette notion ne suffit pas toujours a bien comprendre quels sont les principes de conception du monde SQL qui peuvent être oubliés et ceux qui doivent être appris. Colonne or not colonne ? Prenons un modèle très simple de donnée et essayons de le transposer dans un modèle "orienté colonne": Comme on peut le voir on est passé d'un modèle à 2 dimensions (ligne x colonne) vers un modèle où une valeur est accédée au travers de 2  coordonnées qui sont ici (ligne, colonne) Cette notion de coordonnées est  importante  (c'est pour ça que je la met en gras 2 fois de suite) si l'on veut c

Row Count : HBase Aggregation example

With the coprocessors HBase 0.92 introduces a new way to process data directly on a region server. As a user this is definitively a very exciting feature : now you can easily define your own distributed data services. This post is not intended to help you how to define them (i highly recommend you to watch this presentation if you want to do so) but to quickly presents the new aggregation service shipped with HBase 0.92 that is built upon the endpoint coprocessor framework. 1. Enable AggregationClient coprocessor You have two choices : You can enable aggregation coprocessor on all your tables by adding the following lines to hbase-site.xml : <property> <name>hbase.coprocessor.user.region.classes</name> <value>org.apache.hadoop.hbase.coprocessor.AggregateImplementation</value> </property> or ...you can enable coprocessor only on a table throught the HBase shell : 1. disable the table hbase> disable ' mytable ' 2.

Zookeeper, Netflix Curator and ACLs

If you have one or more Zookeeper "multi-tenant" clusters you may want to protect znodes against unwanted modifications. Here is a very simple and short introduction to the ACL and custom authentication features. This post is not intended to give you best practices about security and Zookeeper, the only goal is to give you a complete example of a custom authentication handler. Complete source code with JUnit test is available here : https://github.com/barkbay/zookeeper-acl-sample/ Use case Let say that your Zookeeper cluster is used by several users. In order to restrict user actions you have decided that each user must prefix all paths with the first letter of his name. User foo is only allowed to create, read, delete and update znodes under the /f znode. User bar is only allowed to create, read, delete and update znodes under the /b znode. Get client authentication data on the server side Zookeeper client authentication can be easily customized , a