Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

dynamic fields and open formula

$
0
0
hi,
I want to divided 2 sum fields, from 2 lignes. the fields is issued by one datasource. I call the field SOURCE in my exemple.

Here is my report structure:
columns ---------- results
FRUITS ---------- PRICE (PRICE issued by sum on group FRUITS) -- CALCULATE
TYPEOF_FRUITS -- PRICE (PRICE issued by sum on group TYPEOF_FRUITS)

I want in a new column (call CALCULATE) : PRICE (TYPEOF_FRUITS) / PRICE (FRUITS)

I try Openformula without succes.
ANy Hints or How to ?
Thanks for the time you spend to me.

Programmatically setting the database interface for a job's child transformations

$
0
0
To accomodate unit testing with HSQLDB, I've been setting the DatabaseInterface for transformations programmatically like this:

Code:

private void adjustDatabase(HasDatabasesInterface hasDatabases, DatabaseMeta databaseMeta) throws TransformationException {
    DatabaseInterface databaseInterface;
    try {
      databaseInterface = DatabaseMeta.getDatabaseInterface(databaseType);
      databaseInterface.setAccessType(DatabaseMeta.TYPE_ACCESS_JNDI);
      databaseInterface.setName(this.databaseName);
      databaseInterface.setDatabaseName("${DB_JNDI_NAME}");
      databaseMeta.setDatabaseInterface(databaseInterface);
      hasDatabases.addOrReplaceDatabase(databaseMeta);
    } catch (KettleDatabaseException e) {
      throw new TransformationException("Error getting database interface " + this.databaseType, e);
    }
  }

This way I can set a variable in my application context and use the Generic interface for HSQLDB testing and the PostgreSQL interface for integration testing or production. So far so good, as long as I'm only trying to do this on transformations. But when I try to run jobs with this paradigm I run into a big problem: I can set the databaseMeta for the JobMeta all right, but when the job is executed the TransMeta objects are created on the fly, and they don't inherit the database properties from the parent job :(. Here's the code in JobEntryTrans:

Code:

  TransMeta transMeta = null;
      switch( specificationMethod ) {
        case FILENAME:
          String filename = space.environmentSubstitute( getFilename() );
          logBasic( "Loading transformation from XML file [" + filename + "]" );
          transMeta = new TransMeta( filename, null, true, this );
          break;

I don't see any way of working the magic here. Does anyone have any ideas how I can tell this instantiated TransMeta to use the Generic database interface? It almost seems like a bug to me...

Distance measure for EM Clustering

$
0
0
Hi!

I have used the EM clustering algorithm provided by Weka 3.6.8 version API. I use the EM class, sets relatively standard options ("-I 100 -N -1 -max 5 -M 1.0E-6 -S 100"), build the clusters (buildClusterer), evaluate them, and are happy with the results returned. Everything is running smoothly.

However, people keep asking me what distance function I have used, and I can not find any information regarding this. So my question is: What distance function (if any), does the EM clusterer provided by Weka use? Are there any options on which distance function to use? And if it does not use a distance function, why?

Hyperlink to a PDF / Excel file.

$
0
0
Hi community,

I'd want to know if it's possible to create an hyperlink within Report Designer which opens a PDF, Excel or any file when I click it. Until now, I've created hyperlinks to another reports and it works fine. But I couldn't get working it with links to other files.
Is it possible?

Thanks in advance.

Cannot start 5.1

$
0
0
I have installed 5.1 with Tomcat 7.0, Java 1.7 and Postgresql 9.2, on Windows 7 x64.

I think I have made all the configutarions correctly, but cannot start.

The first lines of Pentaho.log file:
2014-07-29 14:33:53,901 ERROR [org.pentaho.platform.util.logging.Logger] Error: Pentaho
2014-07-29 14:33:53,902 ERROR [org.pentaho.platform.util.logging.Logger] misc-org.pentaho.platform.engine.core.system.PentahoSystem: org.pentaho.platform.api.engine.PentahoSystemException: PentahoSystem.ERROR_0014 - Error while trying to execute startup sequence for org.pentaho.platform.plugin.services.security.userrole.SecuritySystemListener
org.pentaho.platform.api.engine.PentahoSystemException: org.pentaho.platform.api.engine.PentahoSystemException: PentahoSystem.ERROR_0014 - Error while trying to execute startup sequence for org.pentaho.platform.plugin.services.security.userrole.SecuritySystemListener

...
Caused by: org.pentaho.platform.api.engine.PentahoSystemException: PentahoSystem.ERROR_0014 - Error while trying to execute startup sequence for org.pentaho.platform.plugin.services.security.userrole.SecuritySystemListener
...
Caused by: org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'org.pentaho.platform.engine.core.system.objfac.spring.BeanBuilder#0': FactoryBean threw exception on object creation; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'org.springframework.security.providers.dao.DaoAuthenticationProvider#1' defined in file [C:\pentaho\pentaho-solutions\system\applicationContext-spring-security-jackrabbit.xml]: Cannot create inner bean 'pen:bean#156d12f5' of type
...


Anyone could help me? I've not found anything about these errors.

Excuse me for my bad english.

Regards,
Antonio

FileOutputStream

$
0
0
Please,
I have one problem, i'm try create image to file.

var f1 = new Packages.java.io.FileOutputStream("file.png");
f1.write(imagem.getBytes(), 0, size.getInteger());
f1.close();

I'm viewing the message below:

Javascript error:
Java Script Value 2.0 - Cannot convert org.mozilla.javascript.UniqueTag@3184d05c: NOT_FOUND to byte[]

PSW Role does not bring back data

$
0
0
Hi all,

I created a simple cube(1 dim, 1 fact, 1 hiery) in Schema Workbench 3.6.1 ce and a role called Administrator with 'all' access to the schema, cube, dimension and hierarchy.Published it to the Pentaho BI server 5.0.1 ce (I have cut and paste the xml below). I have Pivot4J and Saiku plug-ins on the server and I can query the cube that I just published but it does not bring back data. What am I missing? Thanks for your help.

DS

<Schema name="Made_Development">
<Cube name="Made_Transactions" visible="true" cache="true" enabled="true">
<Table name="made_trx_history" schema="made" alias="Transactions">
</Table>
<Dimension type="StandardDimension" visible="true" foreignKey="mnoid" highCardinality="false" name="MNO" description="MNO">
<Hierarchy name="MNO_Hiery" visible="true" hasAll="true" allMemberName="allMNO">
<Table name="made_mno" schema="made">
</Table>
<Level name="MNO" visible="true" table="made_mno" column="companyid" type="String" uniqueMembers="true" levelType="Regular" hideMemberIf="Never">
</Level>
<Level name="Master_Account" visible="true" table="made_mno" column="masteraccountnum" type="String" uniqueMembers="false" levelType="Regular" hideMemberIf="Never">
</Level>
</Hierarchy>
</Dimension>
<Measure name="Transaction_Type" column="transaction_type" aggregator="count" visible="true">
</Measure>
</Cube>
<Role name="Administrator">
<SchemaGrant access="all">
<CubeGrant cube="Made_Transactions" access="all">
<DimensionGrant dimension="Measures" access="all">
</DimensionGrant>
<DimensionGrant dimension="MNO" access="all">
</DimensionGrant>
<HierarchyGrant hierarchy="[MNO.MNO_Hiery]" access="all">
</HierarchyGrant>
</CubeGrant>
</SchemaGrant>
</Role>
</Schema>

Creating query to database2 using information from database1

$
0
0
Good afternoon all,
Thank you in advance for any help. I am new to Pentaho and am very pleased to see such a large community.

I am using Pentaho Spoon which has been recently integrated in our environment in partner with Atrium Integrator in our AR environment.


I am creating a new transformation that will:

Pull data from a database based on a specific query(Like name, userid, company)
Then I will need to query another database and get data from that database based on the data in the first database.

so for example, from database1 i will get a big list of names, then i will query another database with those names to get their computer names. Then in the end i will puh the name and computer name to another table.

I am able to get the first database data but am having a hard time figuring out how to query another database based on that info.

Any help is appreciated.
Please move to another thread if needed.

Thank you,
Jeremy Drury
Kettle Spoon 4.1.0

Creating query to database2 using information from database1

$
0
0
Good afternoon all,
Thank you in advance for any help. I am new to Pentaho and am very pleased to see such a large community.


I am using Pentaho Spoon which has been recently integrated in our environment in partner with Atrium Integrator in our AR environment.




I am creating a new transformation that will:


Pull data from a database based on a specific query(Like name, userid, company)
Then I will need to query another database and get data from that database based on the data in the first database.


so for example, from database1 i will get a big list of names, then i will query another database with those names to get their computer names. Then in the end i will puh the name and computer name to another table.


I am able to get the first database data but am having a hard time figuring out how to query another database based on that info.


Any help is appreciated.
Please move to another thread if needed.


Thank you,
Jeremy Drury
Kettle Spoon 4.1.0

Unable to launch Data Integration.app on Mac OS X, but spoon.sh launches fine

$
0
0
I'm having trouble getting the Data Integration.app to launch on Mac OS X. When I launch the app, a dialog pops up complaining "To open Data Integration you need to install the legacy Java SE 6 runtime." Alternatively, I can launch via the spoon.sh launcher script, and Kettle launches just fine with no error messages.

I'd really prefer not to install Java 6, as it's pretty out of date nowadays, but I can't figure out why the Data Integration.app won't launch using one of my two existing JDKs. Any ideas?


Kettle version: Community Edition 5.1.0.0-752
Java version: JDK 1.7.0u60 and JDK 1.8.0u11
Mac OS X version: 10.10

Mondrian XMLA Query Cancelling initiative

$
0
0
At Truven Health Analytics we are currently upgrading our Mondrian version from our older, forked version. One area of concern is Query Cancelling. We currently take advantage of the Cognos support of Microsoft's XMLA query cancel support. However, our current implementation is proprietary. We would like to develop a better implementation which would be suitable for inclusion in the Mondrian XMLA project.


We would like to move forward with the feature request MONDRIAN-689 (http://jira.pentaho.com/browse/MONDRIAN-689). We begin that process with this official call for input. Please leave your input as comments on the above feature request, and not on this message.


A couple problems which must be solved are:


1. Determine which Mondrian queries are part of the associated XMLA request (in house we use query registration).
2. And to quote from Julian: "We'd have to think a bit about session support. The XMLA server uses connections from a pool, so the session id would not map directly onto an id of the back-end mondrian connection. The session concept would therefore have to be introduced in the web layer. We'd have to ensure that connections timeout and go away after some period of time (not hold resources forever) and ensure that session ids are not recycled. My hunch is that we should leverage services from the web container to get session support."


Of course, cancel would be implemented in the XMLA olap4j driver as a secondary part of this initiative.

Design Pattern for recursive Job calls

$
0
0
I have a job that reviews a folder for new files, then processes them. I'd like to make this a continuous process by starting the job, then calling the same job from within itself. In this manner, the job would always be active, but would not start until the last instance has completed. What is the design pattern to do this? I tried simply adding a job step to the job, but received the "Endless loop detected: A Job entry in this job called is calling itself." error. Thanks in advance!

Update where a particular field is null

$
0
0
Hi,

I am a newbie here and have a need where i want to update a table column where the value in any of the rows is null. I have used update with key lookups but don't know how to do the same for null values.
Need help on this.

Split values of a particular value

$
0
0
Hi,

I have a column named score with alpha numeric characters as values example,

Score
90%
20%
90%pass
passed with distinction
Fail
not attempted

I need to extract only columns that has numerical values into another derived column . Expected output ,

Score derv
90% 90%
20% 20%
90%pass 90%
passed with distinction -
Fail
not attempted -

Related to looping or anyother solutions- I am very new.

$
0
0
Hi,

I have simple question.

1. I have a source data with column names like-

phoneno,billing type,timestamp, merchant name
123 ,charge ,yyyymmddhhmmss, M1
123 ,charge ,yyyymmddhhmmss, M1
123 ,charge ,yyyymmddhhmmss, M2


I need to see if a single phone no. has more than 1 billing type as value charge then i need to compare timestamp for both transaction if same then i need to compare merchant name for same, if matches then i will create column name as "stacked".

I need to repeat for each phone number, how do i approach?

I know it is simple, but i am very new to the same.

Thanks,
Ranjan

Deleting BA Reports and Folder through Program

$
0
0
Hi,
I want to delete the reports and folders published in BA server through program instead of doing it manually.
How can I do that?

Regards,
Abhishek

Transformations Won't Start

$
0
0
Hi,

I'm having an issue with getting Transformations to start running.
The Transformations that I'm having the issue with have worked in the past, but I have gone back to add some Filter Rows steps to them to correct some bad data I'm getting out.
Now the Transformations don't run when I press the 'Play' Button in Spoon.
But I can preview the last step and the Transformation will run up to that point.
Below I have copied what I get with the Logging Level set to 'Detailed'.


2014/07/30 17:40:00 - order_out - Transformation is pre-loaded.
2014/07/30 17:40:00 - Spoon - Transformation opened.
2014/07/30 17:40:00 - Spoon - Launching transformation [order_out]...
2014/07/30 17:40:00 - Spoon - Started the transformation execution.
2014/07/30 17:40:00 - order_out - Dispatching started for transformation [order_out]
2014/07/30 17:40:00 - order_out - Nr of arguments detected:0
2014/07/30 17:40:00 - order_out - This is not a replay transformation

At this point the 'Play' Button is reset.

I have found at least one Forum Topic where this has happened before.
But the solution wasn't included.
Any info on what can cause this will be appreciated.

Issue on start up a new pentaho server (5.1)

$
0
0
Hi Team,

Issue on startup of new pentaho server (5.1) . BI server is up but cannot browser solutions tab .? Still loading to show solutions in PUC from more than 2 hours .

Below are my logs :-

org.pentaho.platform.api.repository2.unified.UnifiedRepositoryException: exception while getting file with path "/etc/mondrian"

Reference number: dadafc5b-3c32-42ce-b095-0e2ff446255a
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.callLogThrow(ExceptionLoggingDecorator.java:512)
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.getFile(ExceptionLoggingDecorator.java:260)
at org.pentaho.platform.plugin.action.mondrian.catalog.MondrianCatalogHelper.generateInMemoryDatasourcesXml(MondrianCatalogHelper.java:316)

*******************

Jul 30, 2014 7:02:49 AM org.apache.catalina.core.StandardWrapperValve invoke
SEVERE: Servlet.service() for servlet jaxrsEndpoint-spring threw exception
org.pentaho.platform.api.repository2.unified.UnifiedRepositoryException:
exception while getting tree rooted at path "/"

Reference number: 869267db-e252-4e6f-97ee-de32711734de
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.callLogThrow(ExceptionLoggingDecorator.java:512)
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.getTree(ExceptionLoggingDecorator.java:443)
at org.pentaho.platform.repository2.unified.webservices.DefaultUnifiedRepositoryWebService.getTreeFromRequest(DefaultUnifiedRepositoryWebService.java:160)


*************************

Jul 30, 2014 5:49:12 AM com.sun.jersey.spi.container.ContainerResponse mapMappableContainerException
SEVERE: The RuntimeException could not be mapped to a response, re-throwing to the HTTP container
org.pentaho.platform.api.repository2.unified.UnifiedRepositoryException:
exception while getting file with path "/etc/metadata"

Reference number: 9ad9e779-8609-4351-aa74-8168b895975a
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.callLogThrow(ExceptionLoggingDecorator.java:512)
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.getFile(ExceptionLoggingDecorator.java:260)
at org.pentaho.platform.plugin.services.metadata.PentahoMetadataDomainRepository.getMetadataDir(PentahoMetadataDomainRepository.java:557)
at org.pentaho.platform.plugin.services.metadata.PentahoMetadataDomainRepository.internalReloadDomains(PentahoMetadataDomainRepository.java:446)
at org.pentaho.platform.plugin.services.metadata.PentahoMetadataDomainRepository.getDomainIds(PentahoMetadataDomainRepository.java:342)



Thanks,
Surya Thanuri

Cassandra Input - Data Type

$
0
0
Hi guys!
I need some hints about Cassandra Input step in Pentaho Data Integration.
Let me explain what I've done so far.

----------------QUERY on CASSANDRA CLIENT--------------------------------
I need to take the value from a column family called "Messages", and i'm interested in column1 and value.

For example, we consider two sample lines:
column1="1:asset" value="0x30303a45303a43373a30393a33363a3635" - [valuetipe:HEX] contains a String.
column1="2:metric:Float:temperatureExhaust" value="0x41f00000" - [valuetipe:HEX<arraylist>] contains a float.
(ultimately I will need to encode this information as a string or a float or other types, depending on the need)


----------------QUERY on CASSANDRA INPUT (PENTAHO)----------------
Now, i've tried to reproduce the same query on Cassandra input step in Pentaho Data Integration (QueryMessages.jpg)
and, reading the result, seems that Pentaho deserializing everything as a String.
We consider the two sample lines:

First line
column1="1:asset" and value="0x30303a45303a43373a30393a33363a3635"
correctly decoded as "00:E0:C7:09:36:65" which is the correct string contained in the field.
In this case there are no problem because in the end i need a String.

Second line
column1="2:metric:Float:temperatureExhaust" value="0x41f00000"
Pethaho try to translate from HEX<arraylist> to TEXT, and this generate some problem, because return "A�".

What i try to do
(just for the row which have this structured type)
is to set a Calculator node for coming back to the HEX<arraylist> notation, the same shown in the Cassandra client.

for example
the row in Cassandra
column1="2:metric:Float:temperatureExhaust" value="0x41f00000" (HEX<arraylist>)
in Pentaho become
column1="2:metric:Float:temperatureExhaust" value="A�" (HEX->TEXT)
and using a Calculator the result is
column1="2:metric:Float:temperatureExhaust" value="41FD" (TEXT->HEX)

Now, value="41FD" could be, in such way, related to value="0x41f00000"?
there may be a way to get back to the original information from here?

the ideal solution for me is to use a different deserialization in function of the type indicated in column1.
Is there a way to implement this kind of solution?

If any of you can give me some hints or provide documentation about this topic I'll be grateful !

Grazie!
Federica

How do I receive api call request in job/transformation

$
0
0
Hello,

I have the following situation:

What I do currently:
1. Paste this url into a web browswer (with the correct access codes and dates: http://api.flurry.com/rawData/Sessio...ndDate=ENDDATE

2. Via the browser I receive the following type of response:
{"@reportReady":"false","@type":"Events","@generatedDate":"7/30/14 6:13 AM","@version":"1.0","report":{"@format":"application/json","@reportId":"123498","@reportUri":http://api.flurry.com/rawData/GetReport apiAccessCode=ACCESSCODEBLABLA&reportId=123498}}


3. I paste this url into a browser tab to download the json file: http://api.flurry.com/rawData/GetReport apiAccessCode=ACCESSCODEBLABLA&reportId=123498



I have not been able to figure out how to use PDI to call that url in step one and receive the data in step two. Once I recieve the response data in step 2 i can parse it out, download the json file, process the json file, and load it. I just need to understand how to call the api and receive the response.

Any ideas will be greatly appreciated.

Thanks
Viewing all 16689 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>