Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

Error while running local job from repository

$
0
0
Hello,

I have a problem with my PDI repository. Since I upgraded to Version 6.1 I have following bug when running a local job from my repository job.

I'm using a database repository and have some jobs on the local harddrive of the server that are executed during the process.

I have created a blank repository and recreated the problem. I tested several version of PDI, in Version 5.4 this was possible but since V6 I get this error message.

Here is my logfile
Quote:

2016/06/09 16:33:04 - Job 1 - Start of job execution2016/06/09 16:33:04 - Job 1 - Starting entry [Write To Log]
2016/06/09 16:33:04 - - Repo Job
2016/06/09 16:33:04 - Job 1 - Starting entry [Job]
2016/06/09 16:33:04 - Job - ERROR (version 6.1.0.1-196, build 1 from 2016-04-07 12.08.49 by buildguy) : Error running job entry 'job' :
2016/06/09 16:33:04 - Job - ERROR (version 6.1.0.1-196, build 1 from 2016-04-07 12.08.49 by buildguy) : org.pentaho.di.core.exception.KettleException:
2016/06/09 16:33:04 - Job - Unexpected error during job metadata load
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.run (Job.java:435)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute (Job.java:545)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute (Job.java:864)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute (Job.java:864)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute (Job.java:723)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.entries.job.JobEntryJob.execute (JobEntryJob.java:610)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.entries.job.JobEntryJob.getJobMeta (JobEntryJob.java:1197)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.entries.job.JobEntryJob.getJobMeta (JobEntryJob.java:1219)
2016/06/09 16:33:04 - Job - at org.pentaho.di.repository.kdr.KettleDatabaseRepository.loadJob (KettleDatabaseRepository.java:326)
2016/06/09 16:33:04 - Job - at org.pentaho.di.repository.kdr.delegates.KettleDatabaseRepositoryJobDelegate.loadJobMeta (KettleDatabaseRepositoryJobDelegate.java:327)
2016/06/09 16:33:04 - Job -
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.entries.job.JobEntryJob.getJobMeta(JobEntryJob.java:1291)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.entries.job.JobEntryJob.getJobMeta(JobEntryJob.java:1197)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.entries.job.JobEntryJob.execute(JobEntryJob.java:610)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute(Job.java:723)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute(Job.java:864)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute(Job.java:864)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.execute(Job.java:545)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.Job.run(Job.java:435)
2016/06/09 16:33:04 - Job - Caused by: java.lang.NullPointerException
2016/06/09 16:33:04 - Job - at org.pentaho.di.repository.kdr.delegates.KettleDatabaseRepositoryJobDelegate.loadJobMeta(KettleDatabaseRepositoryJobDelegate.java:327)
2016/06/09 16:33:04 - Job - at org.pentaho.di.repository.kdr.KettleDatabaseRepository.loadJob(KettleDatabaseRepository.java:326)
2016/06/09 16:33:04 - Job - at org.pentaho.di.job.entries.job.JobEntryJob.getJobMeta(JobEntryJob.java:1219)
2016/06/09 16:33:04 - Job - ... 7 more
2016/06/09 16:33:04 - Job 1 - Finished job entry [Job] (result=[false])
2016/06/09 16:33:04 - Job 1 - Finished job entry [Write To Log] (result=[false])
2016/06/09 16:33:04 - Job 1 - Job execution finished
2016/06/09 16:33:04 - Kitchen - Finished!
2016/06/09 16:33:04 - Kitchen - ERROR (version 6.1.0.1-196, build 1 from 2016-04-07 12.08.49 by buildguy) : Finished with errors
2016/06/09 16:33:04 - Kitchen - Start=2016/06/09 16:32:45.120, Stop=2016/06/09 16:33:04.522
2016/06/09 16:33:04 - Kitchen - Processing ended after 19 seconds.
I hope you can help me guys.

Test file input step not working

$
0
0
Hello friends,
I am new to Pentaho technology and started learning recently. I am using Spoon 6.1 version I have difficulty in reading a file using Test file input step. I have a file c:\testfile.txt. when I click Show file content and getting error message stating couldn't find a valid file to work with. Please help me to figure out this issue.

Thanks

Download and process XML File

$
0
0
How can I go about this? I don't see any steps that allow you to download a file.

Thanks!

Distinct-Count not work well

$
0
0
Hello, this is my first post in this community so, i dont know if im in the correct forum.


I have a big problem with a Cube. I define a lot of distinct-Count dimensions and then publish my cube. when i make a query with one of my dimensions, it works properly, but when i add more than one distinc-count dimensions, it return SQL ERROR




After a lot of investigations i find out the problem was in my database engine, becouse it dont allow more than one count(distinct column_name) at the same time.


The question is if there is any way to solve this problem making the cube in other way to make it work?


My database is INFORMIX 11

Running Kettle to Generate Reports and Email

$
0
0
I hope someone knows an easier way than I give below but at least I've found a work around for this though it is far more complicated than I'd wish.

Maybe I need to create a repository in order to run ktr files directly in the PUC, but I would think that it would work right out of the box without
having to build a repository. I instead get the diabolical:

org.pentaho.platform.api.engine.ActionExecutionException: Kettle.ERROR_0004 - The repository specified was not found (null)

So anyway:

Steps for generating and emailing Reports for BISERVER 5.4


This can also be used to apply the path to a subtransformation
executed in biserver. The basis of technique I mostly copied from
Miguel Cunhal, though I had to determine a number of paths on my own.


WARNING: No guarantees on functionality.


This is sort of quick and dirty as I have to get back to work, I just hope
it helps someone.


Note: Replace "reportadmin" directory with whatever name you feel needed.


1. Install RepositorySynchronizer and reboot twice.


2. Synchronize the repository.
---
Server steps
---
3. Enter the ./pentaho-solutions directory.


4. execute: ln -s repositorySynchronizer/public public


5. Create a directory in repositorySynchronizer above named home


6. Create a a directory in repositorySynchronizer/home named reportadmin
---
Developer steps
---
7. Create Kettle transformation using Pentaho Reporting Output as a template.


8. In the step: Create Template and Output paths:
a. Report_Template should = getVariable("Internal.Transformation.Filename.Directory"," ")+"/"+<prpt file>
b. Report_Output should = "./../../pentaho-solutions/home/reportadmin/"+<report name>
(hint: the base directory is biserver-ce/tomcat/bin)
c. Add a Report_Get field to the output of this step.
d. Report_Get should =
"/local/<path to biserver dir>/pentaho-solutions/home/reportadmin/"+<report name>


9. Add an Mail step to the transformation.


10. Use "Report_Get" field as the name of the file to attach.


11. Configure the rest of the Mail step as normal with data fields.


12. Add a Job->Copy Rows to result step. Name this 'output'


13. This is now ready to use in a CDE dashboard.


14. Reminder: If you are creating the report that mails this AFTER this
you'll need to synch the repository before it will work.
---
Scheduling
---
12. Since kettle transforms will not execute in the PUC or Scheduler, this is how
I schedule this.


12a. Here's a twist: You need to edit your Kettle transformation slightly because
OF COURSE quartz uses the tomcat/bin directory as it's base. So go to your transform
and change your Create Template and Output Paths step like so:
Report_Template = "./../../pentaho-solutions/home/reportadmin/"+ "/"+input;

13. Create a generic report in PRD.


14. Go to File->Resources


15. Click on import. Set the Content to text/xml. Name it 'do_emails' or whatever.


16. Make sure you are on the PRD Data Tab for the report (right hand side) and
right click 'Data Sets'. Then select Pentaho Data Integration. I forget if this an advanced
feature, if it is then you need to change the PRD setting to allow for it.


17. Double click on the new datasource, click on the green plus on the left.


18. Name your 'query' whatever.


19. In the file box type: 'do_emails', the box below it should fill up with
the steps in the kettle transformation.


20. Highlight the 'output' step and click OK


21. Save this report as 'gen_email_reports.prpt'.


22. Place this in whatever folder you need to on your BISERVER via Upload
in PUC.


23. Now you can schedule this report to run and it should generate reports
and email them out.

Report Footer page breaks

$
0
0
I'm working on designing an invoice, which includes a tear off section at the bottom of the last page (the report footer section). However, when the height of the section gets a little over two inches, it's splitting it into two pages even though there is enough space to print together. Any ideas on things to check?

I've verified the dynamic-height is set to true, turned off max-height, and avoid-page-break is set to true.

Text File Input previews OK but doesn't output values - examples attached - help!

$
0
0
Hi all,

I am using Pentaho 6.1. I have a strange issue whereby a "text file input" step previews the contents of a file OK, but then the values are not visible in stream.

I have added a "text file output" step to my larger transformation directly after the "text file input" step to show this.

Even stranger, when I copy the same two text input/output steps into their own transformation, the file reads and outputs as expected.

I am very confused here! Can somebody explain to me why this is happening? Or even better, how to fix this read in my larger transformation?

I have attached a zip file containing a copy of the main/larger transformation "test http client - auth token (test).ktr" and also this smaller two-step input/output example "stand-alone text read test.ktr".

Thanks!


Regards,

Chris

EDIT: this also happens when I try running in Pentaho v5.3.
Attached Files

Error when use Split Field step with Apache Kafka Consumer

$
0
0
Hi everyone,


I am using Apache Kafka Consumer to read message from broker. It's ok.
After that, i add a "Split Field" step to split fields in message which is output of Apache Kafka Consumer.
but when i run the transformation, occur error at Split Field step.
the message is bellow:


2016/06/10 10:34:21 - Split Fields.0 - Can only split string fields. [message] is not a string!


please help me to solve this problem.
sorry for my bad english.


many thanks.

Kettle Java API

$
0
0
I am develop a java app which has to integrate with kettle. (multi table input- 1 table output)

I am finding an Example about Java API.
How can i start?

HTTP Post "post a file" not working

$
0
0
Hi Community members!

I am trying to make a webservice call to ibpindex.com following their curl example, listed below

class IBP {
var $filename; //Source filename
var $ibp; //Resoult: JSON Object)
function IBP($filename = false){ //Constructor
if(!empty($filename)) $this->getIBP($filename);
}
function getIBP($filename) {
if(file_exists($filename)) {

//Post fields
$post_data = array();
$post_data['file'] = "@$filename";
$post_data['key'] = 'yourapikey'; // Your api key
//Curl connection
$ch = curl_init();
curl_setopt($ch, CURLOPT_URL, "http://www.ibpindex.com/api/" ); // or "http://www.ibpindex.com/api/index.php"
curl_setopt($ch, CURLOPT_POST, 1 );
curl_setopt($ch, CURLOPT_HEADER, 0);
curl_setopt($ch, CURLOPT_POSTFIELDS, $post_data);
curl_setopt($ch, CURLOPT_RETURNTRANSFER, 1);
$postResult = curl_exec($ch); //return result
if (curl_errno($ch)) {
die(curl_error($ch)); //this stops the execution under a Curl failure
}
curl_close($ch); //close connection
$this->ibp = $postResult;
return $postResult;
}
}
}
# usage: $ibp = new IBP('path/to/file.gpx');
$ibp = new IBP('api_demo_track.gpx');
$ibpindex_json = $ibp->ibp;

I have setup the attached tiny PDI transformation with just 4 steps
- Generate rows, limited to 1, creating two fields, file and key
- Set field value to a constant, setting the key field to my api-key and the file field to the local location of my gpx file
- HTTP Post step, with Request Entity Field set to the file field and the Post a file checkbox checked
- and on the fieds tab both fileds specified as Body parameters, not put in header

But when I run this tranformation a constantly get error reponse from de IBPIndex webservice stating that "No track file received contact www.ibpindex.com"

Can anybody explain why this is not working and what am I doing wrong?

Thnx for you help!
Attached Files

Changing meta after a while

$
0
0
Hi all,

I wonder if Spoon actually supports metadata changing at run-time.

I've implemented a custom plugin that changes the metadata of the outputrow depending on a set of conditions.
This means that processRow() starts with a certain metadata and then, after a while, it changes it.
Of course, the row sent in output is always synchronized with the related metadata.

However, I've noticed that the following steps misbehave (or throw exceptions) because the outputrow does not match anymore the old metadata.
Apparently, most steps assume that the initial metadata won't be changed, and don't check if it has been changed.

I've also tried using a very simple UDJC, and the behavior is still the same.

Can you please confirm me this hypothesis ?

Many thanks.

Possible bug in Unique rows (Hashset)

$
0
0
I had a problem in a more complex scenario and failed with "Unique Rows (Hashset)" processing. So I reduced the problem to a simple test transformation.

Both tests are using a simple string list as input.
test_unique_rows_hash_ok works fine (input comes from a data grid step).
test_unique_rows_hash_failed does not work (input comes from attached testdata.txt CSV file).
Within test_unique_rows_hash_failed "Unique Rows (Hashset)" step does not filter duplicate values which is shown by flag=1 within log for all rows.

I've reproduced results for PDI 6.0 and 6.1 at a Windows 7 machine.

Any ideas?

Best Regards
Martin
Attached Files

Open file on Startup of Spoon - Open With?

$
0
0
I would like to create an "open with" option in windows to open local KTR and KJB files from windows explorer. To do so, I need to pass the file path to spoon via the command line. For kitchen there is a /file. Is there a similar option for spoon? I tried to use /file with spoon.bat but it did not work.

Thanks in advance.

pentaho CDE

$
0
0
Hi,

How can implement the pentaho CDE CCC Line Chart in x axis on slider .


Thanks
Manoj

DELET ROWS - update

$
0
0
Hello,

I have a concern at the integration and updating step.

I have a kettle transformation from which I insert data and I updated my DWH.



but I have fallen in one case, when a Rows deletes from my DB source , even after the launch of update, it remains stored in my database DWH, which reflects on the reliability of my data





can you please help me?



Writing varibles to a log

$
0
0
Hi all,

So I have this:

Get Variables - Name=vCompany, Variable=${Company}
Write to Log - 1 = vCompany

But as you can see from this log extract, the write to log doesn't find it but it is there:

2016/06/11 14:29:47 - Get Variables.0 - field [vCompany] has value [foo]
2016/06/11 14:29:47 - Get Variables.0 - field [vFormattedGetNow] has value [2016-06-11]
2016/06/11 14:29:47 - Get File Names.0 - Finished processing (I=0, O=0, R=0, W=1, U=0, E=0)
2016/06/11 14:29:47 - Get Variables.0 - Finished processing (I=0, O=0, R=1, W=1, U=0, E=0)
2016/06/11 14:29:47 - Write to log.0 - ERROR (version 4.1.0, build 1 from 2013-10-16 15.35.49) : Can not find field [vCompany] in the input stream!
2016/06/11 14:29:47 - Write to log.0 - ERROR (version 4.1.0, build 1 from 2013-10-16 15.35.49) : Unexpected error :
2016/06/11 14:29:47 - Write to log.0 - ERROR (version 4.1.0, build 1 from 2013-10-16 15.35.49) : org.pentaho.di.core.exception.KettleException:

Any help please
Thank you

How kettle import and handle class object of java

$
0
0
I'm a beginner of Kettle,and I face a problem.I need to import the data which is a class object of java,and I know the data type in kettle such as int,string,date,bool are available,and I need to handle the data from class object .
what should i do now.what about using java script or design my own plugins? I do hope some can help me.Thank you very much.:)

Read text file from a specific line

$
0
0
People, need directions .. There goes !!!


You would know tell me if there is any way to read a text file and make the PDI interpret its contents from a particular line? For example: I have a file that has headers with information about those data: title of the table, caption, font, etc. The data table starts effectively from the line 7, and 7 the name of the columns, and the first record from the 8. How can I set this up in the IDP without removing the header before? Thank you!


Hugs!


Wisley Velasco

DateTime Field Out of Range

$
0
0
Hi All,

None of my datetime/timestamp fields have timezones attached, but when running a table input > table output I am getting the following error:

Code:

2016/06/12 14:51:57 - Load into PgSql.0 - Error inserting row into table [order] with values: [24287243423588266772310481094], [26245559582889404658947855782], [5], [2015-02-01 07:10:35], [2015-02-01 10:19:20], [24988991910757856986302626456], [24988991910757856986302626456], [1], [0], [11149120150201], [null], [26217875410359655401500625429], [24474687351930392452477944775], [24474613288252948704694215387], [24474613288252948704694215387], [null], [2015-02-01 07:06:00], [2015-02-01 02:06:00], [2015-02-01 07:00:00], [2015-02-01 02:00:00], [2015-02-01 09:51:00], [2015-02-01 04:51:00], [2015-02-01 09:00:00], [2015-02-01 04:00:00], [null], [2015-02-01 09:45:39], [2015-02-01 04:45:39], [2015-02-01 04:00:00], [2015-02-01 09:00:00], [2015-02-01 10:17:00], [2015-02-01 05:17:00], [2015-02-01 10:00:00], [2015-02-01 05:00:00], [null], [null], [null], [null], [Results], [null], [ResultsComplete], [High], [null], [null], [null], [null], [null], [null], [null], [Normal], [null], [26217875410359655401500625432], [null], [null], [null], [null], [null], [null], [null], [null], [2015-02-01 10:19:19]
2016/06/12 14:51:57 - Load into PgSql.0 -
2016/06/12 14:51:57 - Load into PgSql.0 - Error inserting/updating row
2016/06/12 14:51:57 - Load into PgSql.0 - [Amazon](500310) Invalid operation: date/time field value out of range: "2015-02-31 19:46:00.000000000 +00:00";
2016/06/12 14:51:57 - Load into PgSql.0 -
2016/06/12 14:51:57 - Load into PgSql.0 -
2016/06/12 14:51:57 - Load into PgSql.0 -    at org.pentaho.di.trans.steps.tableoutput.TableOutput.writeToTable(TableOutput.java:377)
2016/06/12 14:51:57 - Load into PgSql.0 -    at org.pentaho.di.trans.steps.tableoutput.TableOutput.processRow(TableOutput.java:118)
2016/06/12 14:51:57 - Load into PgSql.0 -    at org.pentaho.di.trans.step.RunThread.run(RunThread.java:62)
2016/06/12 14:51:57 - Load into PgSql.0 -    at java.lang.Thread.run(Thread.java:745)
2016/06/12 14:51:57 - Load into PgSql.0 - Caused by: org.pentaho.di.core.exception.KettleDatabaseException:
2016/06/12 14:51:57 - Load into PgSql.0 - Error inserting/updating row
2016/06/12 14:51:57 - Load into PgSql.0 - [Amazon](500310) Invalid operation: date/time field value out of range: "2015-02-31 19:46:00.000000000 +00:00";
2016/06/12 14:51:57 - Load into PgSql.0 -
2016/06/12 14:51:57 - Load into PgSql.0 -    at org.pentaho.di.core.database.Database.insertRow(Database.java:1266)
2016/06/12 14:51:57 - Load into PgSql.0 -    at org.pentaho.di.trans.steps.tableoutput.TableOutput.writeToTable(TableOutput.java:255)
2016/06/12 14:51:57 - Load into PgSql.0 -    ... 3 more
2016/06/12 14:51:57 - Load into PgSql.0 - Caused by: java.sql.SQLException: [Amazon](500310) Invalid operation: date/time field value out of range: "2015-02-31 19:46:00.000000000 +00:00";
2016/06/12 14:51:57 - Load into PgSql.0 -    at com.amazon.redshift.client.messages.inbound.ErrorResponse.toErrorException(Unknown Source)

As you can see, the row that threw the error does not contain any dates with timezones, nor are any of the dates on 2015-2-31. Obviously 2015-2-31 is a date that does not exist, so I have no clue where this is coming from. Could a data type mismatch cause an implausible date to attempt to insert? It's almost like it is jumping to what it believes to be the last day of the month, the 31st? My columns in the MySQL source are either timestamp or datetime and they are all timestamp on the postgres destination.

I have been fighting this error all weekend and will take any ideas you have, no matter what they are :)

Apache Kafka Consumer don't get newest data from Broker

$
0
0
Hi everyone,

I am using Apache Kafka Consumer in pentaho data integration to read binary message from broker.
I want to do some real-time dashboard from obtained data.
So, i need the Apache Kafka Consumer to get the newest data from Broker.
My problem is that the Apache Kafka Consumer doesn't read newest message from Broker.
It reads older message.
Please help me to figure out how to configure the Apache Kafka Consumer to solve this problem.
Sorry for my bad english.

Many thanks,
hoang anh.
Viewing all 16689 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>