Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

Publishing cube settings

$
0
0
Good morning,
I've a doubt, i'm trying to publish a cube that i've just created with schema workbench (latest version 6.1).
I've filled correctly Pentaho Credentials, but i don't know what to do with Publish Settings.
Searching online i've read that i had to modify the file locate at: /pentaho-solutions/system/publisher_config.xml but i don't have this file!
I'm using Pentaho bi-server (latest version 6.1).
Can you help me? I' can't figure it out

Thank you

Unable to load old models: backward compatibility problem?

$
0
0
Dear all,

I've been developing for many years now a Java application that uses WEKA as learning library. After upgrading our WEKA jar from 3.7.9 to 3.9.0, we are unable to use the old saved models because the train header can't be properly read anymore. We have identified the problem in the transition from WEKA 3.7.11 to 3.7.12.

These few lines of code can reproduce the error:

import weka.core.SerializationHelper;
import weka.core.Instances;
import hr.irb.fastRandomForest.FastRandomForest;

// read both model and train header
Object[] obj = SerializationHelper.readAll( "classifier 2.0 .model" );
// cast first object to FastRandomForest
FastRandomForest frf = (FastRandomForest) obj[ 0 ];
// cast second object to Instances
Instances data = (Instances) obj[ 1 ];

The exception thrown is the following:

java.lang.NullPointerException
at weka.core.Attribute.enumerateValues(Attribute.java:336)
at weka.core.Attribute.toString(Attribute.java:663)
at java.lang.String.valueOf(String.java:2994)
at java.lang.StringBuffer.append(StringBuffer.java:265)
at weka.core.Instances.toString(Instances.java:1714)
at common.AbstractInterpreter.execute(AbstractInterpreter.java:661)
at common.AbstractInterpreter$ExecuteCode.run(AbstractInterpreter.java:559)

It seems the Instances class has changed fields since version 3.7.11 and some values are not readable any more. Is there any solution to maintain backward compatibility of the models or the users should create brand new models from scratch?

Thanks a lot for your help!

Error connecting to SF sandbox

$
0
0
HI,

could somebody help me, as I'm not able to connect to SF sandboxes (PDI v6.1, Java 8) ? Connecting to => https://test.salesforce.com/services/Soap/u/34.0

Error:
------
Error connecting to Salesforce!; nested exception is:
org.xml.sax.SAXException: Invalid element in com.sforce.soap.partner.fault.UnexpectedErrorFault - upgradeURL

Details:
--------
java.lang.Exception:
Error connecting to Salesforce!
; nested exception is:
org.xml.sax.SAXException: Invalid element in com.sforce.soap.partner.fault.UnexpectedErrorFault - upgradeURL


at org.pentaho.di.ui.trans.steps.salesforceinput.SalesforceInputDialog.test(SalesforceInputDialog.java:1359)
at org.pentaho.di.ui.trans.steps.salesforceinput.SalesforceInputDialog.access$2000(SalesforceInputDialog.java:97)
at org.pentaho.di.ui.trans.steps.salesforceinput.SalesforceInputDialog$22.handleEvent(SalesforceInputDialog.java:1197)
at org.eclipse.swt.widgets.EventTable.sendEvent(Unknown Source)
at org.eclipse.swt.widgets.Widget.sendEvent(Unknown Source)
at org.eclipse.swt.widgets.Display.runDeferredEvents(Unknown Source)
at org.eclipse.swt.widgets.Display.readAndDispatch(Unknown Source)
at org.pentaho.di.ui.trans.steps.salesforceinput.SalesforceInputDialog.open(SalesforceInputDialog.java:1261)
at org.pentaho.di.ui.spoon.delegates.SpoonStepsDelegate.editStep(SpoonStepsDelegate.java:125)
at org.pentaho.di.ui.spoon.Spoon.editStep(Spoon.java:8783)
at org.pentaho.di.ui.spoon.trans.TransGraph.editStep(TransGraph.java:3072)
at org.pentaho.di.ui.spoon.trans.TransGraph.mouseDoubleClick(TransGraph.java:755)
at org.eclipse.swt.widgets.TypedListener.handleEvent(Unknown Source)
at org.eclipse.swt.widgets.EventTable.sendEvent(Unknown Source)
at org.eclipse.swt.widgets.Widget.sendEvent(Unknown Source)
at org.eclipse.swt.widgets.Display.runDeferredEvents(Unknown Source)
at org.eclipse.swt.widgets.Display.readAndDispatch(Unknown Source)
at org.pentaho.di.ui.spoon.Spoon.readAndDispatch(Spoon.java:1347)
at org.pentaho.di.ui.spoon.Spoon.waitForDispose(Spoon.java:7989)
at org.pentaho.di.ui.spoon.Spoon.start(Spoon.java:9269)
at org.pentaho.di.ui.spoon.Spoon.main(Spoon.java:662)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.pentaho.commons.launcher.Launcher.main(Launcher.java:92)

Thanks!

PDI 6.1 and Elasticsearch 2.3.5

$
0
0
Anyone who were able to successfully build a job with PDI 6.1 connecting to Elasticsearch 2.3.5? The default bulk load component does not recognize the elasticsearch server, no matter how the service details are provided.

server address - 127.0.0.1 or http://127.0.0.1 or localhost or http://localhost or whatever and the ports 9200/9300 etc etc.

I doubt if its to do with the lib's that come along with the PDI install that might be causing the conflict.

Pls let me know if anyone were able to get around the connectivity issues.

_Ben

Error in log "ERROR BatchLoader$Batch:1432 - AggGen failed"

$
0
0
Everytime Mondrian is queried about a fact table, the following message is logged:

WARN AggGen:279 - Init: For fact table "FACT_IP_ADDRESS", could not get column name for RolapStar.Column: Fact Count (22): count(null)
ERROR BatchLoader$Batch:1432 - AggGen failed

Why is this the case?

The database tables do not have a column named "FACT_COUNT", instead they have a column named "VALUE".

How to resolve this problem? Everything seems to work fine, but the error continues to be logged.

Data caching

$
0
0
Hi All,

I have managed to store the date and time of the current day inside a variable, which can be used in other transformations and will be used as the filename when I create an XML file. However I have an issue, every time I run the job it saves the previous time in the variable, for instance if I ran the job on 03/10/2016 at 14:44:04 after going through some steps I'd get 03102016144404.xml as the result, and if I run it 30 seconds or even 10 minutes later the result remains the same.

I think it must somehow be caching, but if anyone can provide some help regarding my issue that would be great.

Kind regards.

Adam

Merge Join not executed

$
0
0
I have a transformation making use of Merge Join step. Everything works well when there are actually rows to process. But when the input streams of the Merge Join step "receives" 0 row, it seems that the Merge Join step is never executed. The transformation hangs there for days.
I've read a lot of things regarding large amount of data and circular reference but I could not find anything clearly addressing the case where there are no rows to process.
The logs show that the two steps feeding the Merge Join step officially "Finished processing" and "Signalled output done" to the Merge Join step. After that ... nothing!
I'm using Kettle 4.3 (I know we have to migrate to a more recent version... it's in the pipe but I need to fix this now!).

FYI: The job is ran periodically to update some calculations and it can happen that nothing new has happened in the database and consequently there is no row requiring an update.
Thanks for any hint.

ParalledPeriod with parameter for period

$
0
0
Hi all, I want to do a consultation with ParalledPeriod. The difficulty I find that the number of periods must be variable and must be passed as a parameter, I can not find how. Can somebody help me?. If not could, how I can simulate this operation ?.
Thank you

Reading all files from multiple sub-folders

$
0
0
I need to be able to read all files from a root folder and all of it's sub directories under it. Is there a way to do this? I tried using the GetFolders step to pass that list to Get File Names and that did not work. I want to get everything that is not a folder under that root directory.

Thanks for the help.

Start of Job Execution Slow

$
0
0
I have read the posts for the cfgbuilder - warning that will be addressed in v7, but I want to know if that is making my job start later then normal. Below is a copy of my job output for a normal run:
2016/10/03 09:29:04 - Kitchen - Start of run.
2016/10/03 09:29:04 - RepositoriesMeta - Reading repositories XML file: /home/pentaho-pw/.kettle/repositories.xml
2016/10/03 09:29:05 - cfgbuilder - Warning: The configuration parameter [org] is not supported by the default configuration builder
for scheme: sftp
2016/10/03 09:29:12 - pw_stat_flattened_5_mins_ctrlStart - Start of job execution

2016/10/03 09:29:14 - Kitchen - Start=2016/10/03 09:29:04.315, Stop=2016/10/03 09:29:14.855
2016/10/03 09:29:14 - Kitchen - Processing ended after 10 seconds.

If you see the cfgbuilder -Warning and then the Start of job execution is only 7 seconds.
This job executes most of the time in 8 to 14 seconds from start to finish. Periodically I have the delay that you see below.

2016/10/03 09:33:05 - Kitchen - Start of run.
2016/10/03 09:33:05 - RepositoriesMeta - Reading repositories XML file: /home/pentaho-pw/.kettle/repositories.xml
2016/10/03 09:33:07 - cfgbuilder - Warning: The configuration parameter [org] is not supported by the default configuration builder
for scheme: sftp
2016/10/03 09:35:15 - pw_stat_flattened_5_mins_ctrlStart - Start of job execution

If you see the cfgbuilder -Warning and then the Start of job execution there is a 2mins and 8secs.
This job executes most of the time in 8 to 14 seconds from start to finish as you see the job execution at the top.

Does any one know what is causing this delay?

Dimension Update using SQLite

$
0
0
Hi

I'm trying how to populate data using SCD Type 2.

I created input source using text file and output table using SQLite.

Input source contains:
- customer_id
- customer_name
- customer_address
- customer_city

Output table in SQLite:
- customer_dim_id
- version
- date_from
- date_to
- customer_id
- customer_name
- customer_address
- customer_city



Screenshot from 2016-10-04 10-11-35.png

When I ran the transformation, it raised errors.
Customer Dimension.0 - ERROR (version 6.1.0.1-196, build 1 from 2016-04-07 12.08.49 by buildguy) : Because of an error this step can't continue:
Customer Dimension.0 - Unexpected conversion error while converting value [customer_dim_id String] to an Integer
Customer Dimension.0 - java.lang.Long cannot be cast to java.lang.String

I just follow the instruction/tutorial from the web, in the original version, they used H2 database server as a target output. I changed it to SQLite, because I don't have the H2 database server.

Need your help how to test / simulated the process using SQLite.


Regards
Attached Images

Can Someone Explain about Insert/update in PDI with as simple example?

$
0
0
Hi,

I started newly working on PDI,as I'm going through the components and found so many issues.
Can Someone Explain about Insert/update in PDI with as simple example?


Regards,
PrudhviTeja

How to pass variables from the application url to table input

$
0
0
Hi,
I am accessing pentaho report from my web application using
localhost:8080/pentaho/api/repos/:PentahoPublishReports:testReport.prpt/report?empId=3501&output-target=table/html;page-mode=stream&dashboard-mode=true
where empId has a dynamic value and testReport.prpt uses data integration file to query the database.

Data integration has the below transformation steps:
TableInput ---->Text file output---> sort rows.

TableInput has the query "select * from data where id=${empId}"

Please let me know how to pass the empId parameter from my application url to data integration table input.

Thanks,
Padma Priya N.

Create a new kettle database repository with command line

$
0
0
HI,
I use kettle 4.4
How to create a new kettle database repository with command line
Thanks

Parameters in Penrtaho and Multiple Job Instance

$
0
0
Hi All,

Is it possible to pass the parameter file to a pentaho job and define the parameter for all the jobs in the parameter file like Informatica?. Also can we run the same Job with multiple instances in pentaho.?

Regards,
Amit Terse

Insert/Update is only creating one row

$
0
0
Quote:

Originally Posted by Suhendri View Post
Hi PrudhviTeja,

Maybe this simple example about insert update will help you.

1. I created a text file for input. I named it Customer. With 4 fields. Customer_ID, Customer_Name, Customer_Address and Customer_City. I used semi colon as separator/delimiter for each fields. And create some records in it.
2. Create a new transformation
3. Add CSV Input File to the new transformation. Edit all the necessary or required information, such as file name, delimiter, and fields information (You can use get Fields to retrieve field information). Try to preview the data
4. Add Insert/Update. Edit the required information like connection, target table, look up value and update fields
5. Run the transformation

Attachment 17194

Edit the text file, change some field content, and run again the transformation to see the updated results.

Perhaps this simple example will help you.

I am using the insert/update step and for some reason it only adds in one row, not multiple ones. I am getting my data from the "get file names" step.

Hive Driver Fetch Size

$
0
0
Can someone tell me why the Hive Driver is only fetching 50 record chunks?

Is there a way to request larger chunks ... 5000, 10000?

CDE Dashboard export - Dynamic title not being exported

$
0
0
Hi All,

I am working on a CDE dashboard in which we are setting the title dynamically in pre-execution function.
we have created an export button component as well (using CGG), but the title is not being shown when we export the chart.
I have tried to set the title in post fetch as given in,
http://redmine.webdetails.org/projec..._Tips_&_Tricks

but still not getting it in export.
Need help.

How will be the career in Pentaho ?

$
0
0
Hello,
I heard that Pentaho is on demand in the market. I would like to know how Pentaho will be helpful in accelerating my career and what are the prerequisites for Pentaho BI course.

Thanks in advance.

s3db file export

$
0
0
Hi All,

I have a s3db file and i want all the data to be migrated to a staging database table(mysql).

Any idea how to use s3db file input ??

Thanks in advance
Chinmay
Viewing all 16689 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>