Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

Mondrian 4 release date?

$
0
0
Hi,
Single question :-) : What is the expected release date of Mondrian 4?

gr. Martin

Excel Input - XLSX - POI can't find file in network

$
0
0
sooo... hello everyone, how are you all doing? After a brief time away from here, I come back with a huge problem - Excel Input, when set to use Apache POI for XLSX files, can't seem to find files in network. I'm using ${Internal.Transformation.Filename.Directory} and such other variables to set the filename on it - if I'm set on a XLSX file, I get something like this:

Code:

2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - ERROR (version 5.0.1-stable, build 1 from 2013-11-15_16-08-58 by buildguy) : Error processing row from Excel file [file:////df5008fs201/GECAP/GECAP08_INTEGRACAO_RISCOS/Dados/CAPITAL/DLO/2013/2013-10/DRG 2013-10 2041.xlsx] : org.pentaho.di.core.exception.KettleException: 2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - java.io.FileNotFoundException: file:\df5008fs201\GECAP\GECAP08_INTEGRACAO_RISCOS\Dados\CAPITAL\DLO\2013\2013-10\DRG 2013-10 2041.xlsx
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - file:\df5008fs201\GECAP\GECAP08_INTEGRACAO_RISCOS\Dados\CAPITAL\DLO\2013\2013-10\DRG 2013-10 2041.xlsx
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - ERROR (version 5.0.1-stable, build 1 from 2013-11-15_16-08-58 by buildguy) : org.pentaho.di.core.exception.KettleException:
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - java.io.FileNotFoundException: file:\df5008fs201\GECAP\GECAP08_INTEGRACAO_RISCOS\Dados\CAPITAL\DLO\2013\2013-10\DRG 2013-10 2041.xlsx
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - file:\df5008fs201\GECAP\GECAP08_INTEGRACAO_RISCOS\Dados\CAPITAL\DLO\2013\2013-10\DRG 2013-10 2041.xlsx
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at org.pentaho.di.trans.steps.excelinput.poi.PoiWorkbook.<init>(PoiWorkbook.java:78)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at org.pentaho.di.trans.steps.excelinput.WorkbookFactory.getWorkbook(WorkbookFactory.java:38)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at org.pentaho.di.trans.steps.excelinput.ExcelInput.getRowFromWorkbooks(ExcelInput.java:603)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at org.pentaho.di.trans.steps.excelinput.ExcelInput.processRow(ExcelInput.java:471)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at org.pentaho.di.trans.step.RunThread.run(RunThread.java:60)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at java.lang.Thread.run(Unknown Source)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - Caused by: java.io.FileNotFoundException: file:\df5008fs201\GECAP\GECAP08_INTEGRACAO_RISCOS\Dados\CAPITAL\DLO\2013\2013-10\DRG 2013-10 2041.xlsx
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at org.apache.poi.ss.usermodel.WorkbookFactory.create(WorkbookFactory.java:83)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    at org.pentaho.di.trans.steps.excelinput.poi.PoiWorkbook.<init>(PoiWorkbook.java:70)
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 -    ... 5 more
2014/01/07 17:15:09 - Input: RWACPAD-DRG.0 - Finished processing (I=0, O=0, R=0, W=0, U=0, E=1)

From the poor knowledge I have on this, it seems POI's parsing the filename's path erroneously - changing \\ to \ and making it not look at the network. I've already tried replacing PDI 5's POI 3.9 with the 3.10 Beta2 but it also didn't work. Any ideas?

Funny thing - on PDI 4.4 and POI 3.8, it works...

two parameters from one query?

$
0
0
Hey PRD experts,

Is it possible to run a query on an external data source, and then query the resulting table multiple times to populate multiple parameters?

I have a query that returns two columns like the table at the bottom of this post. Using this query as a source, I would like to create a [day_param] that displays distinct values from the first column (e.g., SELECT DISTINCT day). And I would like to create a [mkt_param] that displays values from the second column based on the user's [day_param] selection (e.g., SELECT mkt WHERE day = [day_param]).

I am trying to avoid querying my external data source twice. Is this possible?

Thanks,
RR

day
mkt
2013-12-21
CA
2013-12-21
CN
2013-12-21
DE
2013-12-14
CA
2013-12-14
CN
2013-12-14
DE
2013-12-14
ES
2013-11-30
CA
2013-11-30
CN
2013-11-30
DE
2013-11-30
FR

Trouble connecting to mongo

$
0
0
Hi,

New to PDR.

I am trying to connect to mongo db using PDR 5.0. I have the mongo-java-driver-2.11.3.jar in the \report-designer\lib directory but I still get the error 'Unable to retrieve database names' when I try to get a list of databases.

I restarted the app after I installed to driver and I definatly have the correct user/pw.

Any ideas? Do I have to set a path?

Thanks,

Database lookup fails if Database Insert is running in the same transformation

$
0
0
PDI: 4.4.0-stable

I have a MySQL table (personids) that I update while running my transformation using Table Output. It has

person_id
person_index
customer_id

The table updates successfully. I can run view it using the MySQL shell to see the tuples that I wrote.

While the same transformation is running, I do subsequent Database Value lookups on the same table using the customer_id as the table field so I can extract the person_id.

The problem is that my Database lookup is failing to see that the personids table is already populated so I am not getting any valid person_id. If I stop the transformation and rerun it without deleting my personids table though, my lookup succeeds.

Any thoughts?
Thanks,
-phylanx

Kettle repository file entries seem to override kettle command line prompts

$
0
0
Hello,
I've created a repository in my database using spoon. All works fine.
However I am trying to run a command line kettle script as below:

./kitchen.sh /user:"PDI_REPOSITORY" /pass:"entahox" /rep:"MarcosAWSDW1" /dir:"/" /job:"TEST_TEMP_JOB"

This runs fine but if I want to run this using a different repository for test the command line arguments except the job parameter are ignored and the job runs on the default repository. Also, if I enter an invalid user id or password on the command line the job continues to run successfully. It's almost as though these command parameters are never used and kitchen.sh only uses the job parameter.

What am I doing wrong? Help!

Any guidance is much appreciated,
John Byrne

How to process a Kettle transformation once per filename

$
0
0
Hello all,

I'm trying to read multiple XML files from a folder, to compile all the data they have (all of them have the same XML structure), and than save that data MySql Database.


I have already designed the transformation for reading XML files and process once per file using following link:

http://www.timbert.net/doku.php?id=t...ansonceperfile


I need to move the XML files which are already processed or write some logic so that already processed XML are no need to processed again.


It would be really helpful if someone provide some useful suggestion.


Thanks in advance, looking forward to reply.

Issue While connecting to Hortonworks Sandbox 2.0 From Pentaho Trial Edition 5.0.2

$
0
0
Hello All,
I have just started evaluating the Pentoho for our Business Integration purpose. We are having a Hortonworks Sandbox 2.0 Environment which is up and running successfully.

I Created a job to copy a file from local disk to HDFS. When I run the job, it's failing.

The logger prints the below exception.

Quote:

2014/01/08 11:32:24 - Hadoop Copy Files - ERROR (version 5.0.2, build 1 from 2013-12-04_15-52-25 by buildguy) : Can not copy file/folder [file:///E:/Data/FifteenGigaText.txt] to [hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test]. Exception : [
2014/01/08 11:32:24 - Hadoop Copy Files -
2014/01/08 11:32:24 - Hadoop Copy Files - Unable to get VFS File object for filename 'hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test' : Could not resolve file "hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test".
2014/01/08 11:32:24 - Hadoop Copy Files -
2014/01/08 11:32:24 - Hadoop Copy Files - ]
2014/01/08 11:32:24 - Hadoop Copy Files - ERROR (version 5.0.2, build 1 from 2013-12-04_15-52-25 by buildguy) : org.pentaho.di.core.exception.KettleFileException:
2014/01/08 11:32:24 - Hadoop Copy Files -
2014/01/08 11:32:24 - Hadoop Copy Files - Unable to get VFS File object for filename 'hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test' : Could not resolve file "hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test".
2014/01/08 11:32:24 - Hadoop Copy Files -
2014/01/08 11:32:24 - Hadoop Copy Files -
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.core.vfs.KettleVFS.getFileObject(KettleVFS.java:162)
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.core.vfs.KettleVFS.getFileObject(KettleVFS.java:105)
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.entries.copyfiles.JobEntryCopyFiles.ProcessFileFolder(JobEntryCopyFiles.java:378)
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.entries.copyfiles.JobEntryCopyFiles.execute(JobEntryCopyFiles.java:326)
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.execute(Job.java:678)
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.execute(Job.java:815)
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.execute(Job.java:500)
2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.run(Job.java:407)
2014/01/08 11:32:24 - HadoopFileCopy - Finished job entry [Hadoop Copy Files] (result=[false])
2014/01/08 11:32:24 - HadoopFileCopy - Job execution finished
2014/01/08 11:32:24 - Spoon - Job has ended.

When i try to google it up, I landed on the page where pentaho talked about the various hadoop distros configuration and compatibility with PDI.

http://wiki.pentaho.com/display/BAD/...ro+and+Version

In this page, it's clearly stated that HDP 2.x is not supported by Pentaho unless we make some manual effort.

Update : I found a Jira http://jira.pentaho.com/browse/PDI-10807 Ticket available for the same reason what I 've asked here for.... The ticket status states it is closed. But no idea what it talks , and what i have to do......

Can Any one guide me through this to get it done.

Thanks In Advance

Regards,
VAP.

Bug with Log4j logging?

$
0
0
Hi ,

We are planning to upgrade our kettle to 5.0.1 from 4.0.1 . But facing some issues with the logging.

We are redirecting the logging to log4j logger.

In PDI-5 , we are not getting any log entries using the same log4j.xml file as of earlier versions.

Here is our log4j.xml file which works fine in earlier version but not in PDI-5 :

<!DOCTYPE log4j:configuration SYSTEM "log4j.dtd">
<log4j:configuration xmlns:log4j="http://jakarta.apache.org/log4j/" debug="false">

<!-- ============================== -->
<!-- Append messages to the console -->
<!-- ============================== -->

<appender name="CONSOLE" class="org.apache.log4j.ConsoleAppender">
<param name="Threshold" value="ERROR"/>
<param name="Target" value="System.out"/>

<layout class="org.apache.log4j.PatternLayout">
<!-- The default pattern: Date Priority [Category] Message\n -->
<param name="ConversionPattern" value="%d{ABSOLUTE} %-5p [%c{1}] %m%n"/>
</layout>
</appender>

<appender name="ETLJOB" class="org.apache.log4j.FileAppender">
<param name="File" value="/home/pradeepm/etl.log"/>

<param name="Append" value="true"/>


<param name="ImmediateFlush" value="true"/>
<param name="BufferedIO" value="false"/>

<layout class="org.apache.log4j.PatternLayout">
<!-- Extended pattern: Date Priority [Category][SessionID] Message\n -->
<param name="ConversionPattern" value="%d %-5p [%t] %m%n"/>
</layout>
</appender>


<category name="org.pentaho.di" additivity="false">
<priority value="debug"/>
<appender-ref ref="ETLJOB"/>
</category>

<!-- ======================= -->
<!-- Setup the Root category -->
<!-- ======================= -->
<root>
<appender-ref ref="CONSOLE"/>
</root>

</log4j:configuration>

Can anyone please help us here ?

I have been trying to find a solution for the past few days with no success.

CST for Pentaho 5

$
0
0
Does anybody know release date or some hints about Community Startup Tabs project for Pentaho 5?

PDI 5.0.1: Can't read log in slave server monitor from remotely executed jobs

$
0
0
Hi,

I am running PDI 5.0.1 Carte on a Windows 2008 server and the jobs are executed remotely with PDI 5.0.1 on a Windows 7 PC. The jobs and transformations are stored in a file repository.
When I am looking at a running job in the slave server monitor, I only get this message:

Code:

Unable to decode logging from remote server : java.util.zip.ZipException: Not in GZIP format
java.util.zip.ZipException: Not in GZIP format
    at java.util.zip.GZIPInputStream.readHeader(Unknown Source)
    at java.util.zip.GZIPInputStream.<init>(Unknown Source)
    at java.util.zip.GZIPInputStream.<init>(Unknown Source)
    at org.pentaho.di.cluster.HttpUtil.decodeBase64ZippedString(HttpUtil.java:140)
    at org.pentaho.di.www.SlaveServerJobStatus.<init>(SlaveServerJobStatus.java:110)
    at org.pentaho.di.www.SlaveServerJobStatus.fromXML(SlaveServerJobStatus.java:136)
    at org.pentaho.di.cluster.SlaveServer.getJobStatus(SlaveServer.java:716)
    at org.pentaho.di.ui.spoon.SpoonSlave.treeItemSelected(SpoonSlave.java:445)
    at org.pentaho.di.ui.spoon.SpoonSlave$1.widgetSelected(SpoonSlave.java:265)
    at org.eclipse.swt.widgets.TypedListener.handleEvent(Unknown Source)
    at org.eclipse.swt.widgets.EventTable.sendEvent(Unknown Source)
    at org.eclipse.swt.widgets.Widget.sendEvent(Unknown Source)
    at org.eclipse.swt.widgets.Display.runDeferredEvents(Unknown Source)
    at org.eclipse.swt.widgets.Display.readAndDispatch(Unknown Source)
    at org.pentaho.di.ui.spoon.Spoon.readAndDispatch(Spoon.java:1227)
    at org.pentaho.di.ui.spoon.Spoon.waitForDispose(Spoon.java:7368)
    at org.pentaho.di.ui.spoon.Spoon.start(Spoon.java:8673)
    at org.pentaho.di.ui.spoon.Spoon.main(Spoon.java:625)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
    at java.lang.reflect.Method.invoke(Unknown Source)
    at org.pentaho.commons.launcher.Launcher.main(Launcher.java:134)

I have found these two JIRA issues which seem to describe my problem too:

http://jira.pentaho.com/browse/PDI-10309
http://jira.pentaho.com/browse/PDI-9900

Does anyone here got an idea, what the reasons for this error could be?
Thanks for any hint!

Bobse

Japanese font cannot display

$
0
0
Dear all,

I have a problem about character encoding(I think).

The problem is:
I have crated the execute prpt component in CDE dashboard and it can display japanese words on browser.(Please refer to the attached img1.jpg)
but after I download this data as CSV, and open on browser, it cannot display japanese words.(Please refer to the attached img2.jpg)

Could any one please guide me how can I do to be available japanese word when open with excel????

Thanks all..
Attached Images

Problem while creating MDX query

$
0
0
Hello guys

I have a requirement to create generate data based on 3 dimensions with measures on it. e.g

i have dimensions like country,state,product and want to display my data like
dimesnion measures
country 100
state 200
product 300

Any help would be appreciated

Is there any way to make a calendar view in c-tools?

$
0
0
Hi:


We want to make a dashboard where appears a calendar chart. Is there any component or way to do it with ctools?


We installed viz-calendar plug-ing in Pentaho. How we can use it from ctools?


And a last question: We can use/import d3.js libraries in c-tools?


Thanks.

Help needed in setting Kettle Clustering

$
0
0
Hello Everone,

I want to set up Clustering in Kettle . I am using Version 5.1.
When i google it, i got an idea how to set up, but in all the documents ,the database connections are hard coded.
My question is
"How can i use jndi Properties" in setting the clustering environment in kettle

unable to start pentaho BI server

$
0
0
Hi all,

Got below mentioned error ,while starting BI server


05:38:55,111 ERROR [RepositoryImpl] Failed to initialize workspace 'default'
javax.jcr.RepositoryException: Unable to initialize query handler: org.apache.jackrabbit.core.query.lucene.SearchIndex@766ff867
at org.apache.jackrabbit.core.config.RepositoryConfigurationParser$1.getQueryHandler(RepositoryConfigurationParser.java:658)
at org.apache.jackrabbit.core.config.RepositoryConfig.getQueryHandler(RepositoryConfig.java:1037)
at org.apache.jackrabbit.core.SearchManager.<init>(SearchManager.java:171)
at org.apache.jackrabbit.core.RepositoryImpl.getSystemSearchManager(RepositoryImpl.java:614)
*************
*********
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:289)
at org.apache.catalina.startup.Bootstrap.main(Bootstrap.java:414)
Caused by: java.io.FileNotFoundException: /usr/local/pentaho/server/biserver-ee/pentaho-solutions/system/jackrabbit/repository/repository/index/_nk/_2.prx (No such file or directory)
at java.io.RandomAccessFile.open(Native Method)
at java.io.RandomAccessFile.<init>(Unknown Source)


i lost _2.prx after re-start of my machine , can some one guide me how can i make my bi server up

pentaho 5.0 EE

Thanks,
Surya thanuri


YAMFork (Yet Another Mondrian 4 Question)

$
0
0
We are running a forked version of Mondrian 3.2. We have heavily modified XMLA Servlet to provide XML streaming and specialized security conections to our business reporting software.

We are beginning a new version of our product and would like to upgrade Mondrian. We like the fact that XMLA and OLAP4J will be separated from the other code in Mondrian 4. We believe this would give us the option of just working on migrating the XMLA servlet to hopefully ease the pain and get us on the latest technology.

However, when I research what is going on with Mondrian 4, the only development notes I see here are a year old. Has development been going on, and if so, where is it available?

Also, does our migration plan sound feasible.

And, of course (you saw this coming), when will 4 be released?

Views instead of Tables

$
0
0
Hi,
I'm trying to create a cube based on a sql view.
Can someone give me a sample about how to do it?
I just put the code in the cdata field. Just the Select ...... From.... Group By.....
Is that ok?
And how do I define a measure?
I mean, I tried but I couln't find the column, how this will appear?

Thanks in advance.

How to log a raw get transaction from a http client?

$
0
0
Hello PDI Lovers,

I need to log what exactly is the http client is transacting.

Can anybody tell me how?!

Thanks,
Holger

Release notes for 3.6

$
0
0
Does anyone have the release notes and change log for 3.6, they don't seem to be out there with the distribution.
Viewing all 16689 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>