Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

is it possible to use project from EE version in CE version as well

$
0
0
Hi All,

I have used Pentaho EE edition form my project and I have realized that my license is expired for this version. Since I am not planning to buy EE version I am trying to use the same project files from the CE edition then I am getting lot of errors.

Example error: Unable to load job info from XML node
Index: 17, Size: 17

Is it possible to use the same Project from EE edition in CE edition as well..

Regards,
Rama

Pentaho issue on execute start-pentaho.bat

$
0
0
Hi Everyone,

I'm having an issue on trying to startup the bi-server.
The server always worked fine, but some days ago, when I try to start (executing the file start-pentaho.bat) the console shows only those lines on file catalina.2016-01-04.log (folder tomcat/logs):

Jan 04, 2016 3:03:44 PM org.apache.catalina.core.AprLifecycleListener init
INFO: Loaded APR based Apache Tomcat Native library 1.1.32 using APR version 1.5.1.
Jan 04, 2016 3:03:44 PM org.apache.catalina.core.AprLifecycleListener init
INFO: APR capabilities: IPv6 [true], sendfile [true], accept filters [false], random [true].
Jan 04, 2016 3:03:45 PM org.apache.catalina.core.AprLifecycleListener initializeSSL
INFO: OpenSSL successfully initialized with version OpenSSL 1.0.1j 15 Oct 2014
Jan 04, 2016 3:03:46 PM org.apache.coyote.http11.Http11AprProtocol init
INFO: Initializing Coyote HTTP/1.1 on http-8070
Jan 04, 2016 3:03:46 PM org.apache.coyote.ajp.AjpAprProtocol init
INFO: Initializing Coyote AJP/1.3 on ajp-9009
Jan 04, 2016 3:03:46 PM org.apache.catalina.startup.Catalina load
INFO: Initialization processed in 2390 ms
Jan 04, 2016 3:03:46 PM org.apache.catalina.core.StandardService start
INFO: Starting service Catalina
Jan 04, 2016 3:03:46 PM org.apache.catalina.core.StandardEngine start
INFO: Starting Servlet Engine: Apache Tomcat/6.0.43
Jan 04, 2016 3:03:46 PM org.apache.catalina.startup.HostConfig deployDescriptor
INFO: Deploying configuration descriptor pentaho.xml
Jan 04, 2016 3:03:51 PM org.apache.catalina.startup.HostConfig deployDirectory
INFO: Deploying web application directory pentaho-style
Jan 04, 2016 3:03:51 PM org.apache.catalina.startup.HostConfig deployDirectory
INFO: Deploying web application directory ROOT
Jan 04, 2016 3:03:51 PM org.apache.catalina.startup.HostConfig deployDirectory
INFO: Deploying web application directory sw-style
Jan 04, 2016 3:03:51 PM org.apache.coyote.http11.Http11AprProtocol start
INFO: Starting Coyote HTTP/1.1 on http-8070
Jan 04, 2016 3:03:51 PM org.apache.coyote.ajp.AjpAprProtocol start
INFO: Starting Coyote AJP/1.3 on ajp-9009
Jan 04, 2016 3:03:51 PM org.apache.catalina.startup.Catalina start
INFO: Server startup in 5351 ms

And just stop. Anything happens then, and the server dosen't start.
Does anyone, have already face it?
The other logs files are empty.

I didn't changed any port, any file, nothing.
Any clue will help me.

Thanks, and sorry any english spelling mistake. hehe

Cannot run jobs remotely (on a server) from a client machine using Carte

$
0
0
Hello,
I have been trying to setup PD! 6.X CE as client server environment using carte (so that users can run their jobs on a central server). I used the following post as a reference: http://wiki.pentaho.com/display/EAI/...le+environment. So far I have been able to 1) Successfully run Carte on my server and check its status.
2) Setup a repository in MySQL and connect to it (both from server and client machine)
3) Create a test job and setup the slave server settings
When I launch the job (with the remote execution option) from a client machine the connection times out and I get the below error:

org.pentaho.di.core.exception.KettleException:
java.net.ConnectException: Connection timed out: connect
Connection timed out: connect


at org.pentaho.di.trans.Trans.sendToSlaveServer(Trans.java:4298)
at org.pentaho.di.ui.spoon.delegates.SpoonTransformationDelegate.executeTransformation(SpoonTransformationDelegate.java:891)
at org.pentaho.di.ui.spoon.Spoon$31$1.run(Spoon.java:8468)
at org.eclipse.swt.widgets.RunnableLock.run(Unknown Source)
at org.eclipse.swt.widgets.Synchronizer.runAsyncMessages(Unknown Source)
at org.eclipse.swt.widgets.Display.runAsyncMessages(Unknown Source)
at org.eclipse.swt.widgets.Display.readAndDispatch(Unknown Source)
at org.pentaho.di.ui.spoon.Spoon.readAndDispatch(Spoon.java:1339)
at org.pentaho.di.ui.spoon.Spoon.waitForDispose(Spoon.java:7939)
at org.pentaho.di.ui.spoon.Spoon.start(Spoon.java:9214)
at org.pentaho.di.ui.spoon.Spoon.main(Spoon.java:653)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.pentaho.commons.launcher.Launcher.main(Launcher.java:92)
Caused by: java.net.ConnectException: Connection timed out: connect
at java.net.DualStackPlainSocketImpl.connect0(Native Method)
at java.net.DualStackPlainSocketImpl.socketConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.doConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.connectToAddress(Unknown Source)
at java.net.AbstractPlainSocketImpl.connect(Unknown Source)
at java.net.PlainSocketImpl.connect(Unknown Source)
at java.net.SocksSocketImpl.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at java.net.Socket.<init>(Unknown Source)
at java.net.Socket.<init>(Unknown Source)
at org.apache.commons.httpclient.protocol.DefaultProtocolSocketFactory.createSocket(DefaultProtocolSocketFactory.java:80)
at org.apache.commons.httpclient.protocol.DefaultProtocolSocketFactory.createSocket(DefaultProtocolSocketFactory.java:122)
at org.apache.commons.httpclient.HttpConnection.open(HttpConnection.java:707)
at org.apache.commons.httpclient.MultiThreadedHttpConnectionManager$HttpConnectionAdapter.open(MultiThreadedHttpConnectionManager.java:1361)
at org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:387)
at org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171)
at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397)
at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323)
at org.pentaho.di.cluster.SlaveServer.sendXML(SlaveServer.java:483)
at org.pentaho.di.trans.Trans.sendToSlaveServer(Trans.java:4261)
... 15 more
How can i solve this issue and run jobs remotely? (I am able to ping the server machine from my client)
Thank you,
Ash

Pentaho Business Analytics

$
0
0
Hi,

I am new to pentaho. Can someone help me with Pentaho download for 32 bit OS??

I could find only 64 bit version download in the website? Any alternative how do i use it in 32 bit system.


Regards
Alben

HTTP Status 404 - /pentaho/Home- Error

$
0
0
Hello,
I am using Pentaho 5.4 (enterprise) , when i am trying to start the user console (http://localhost:8080), the following message appears:

HTTP Status 404 - /pentaho/Home


type Status report
message /pentaho/Home
description The requested resource is not available.
Apache Tomcat/6.0.43

i was defining JNDI connections for the BA server- Tomcat JNDI connections

in the attachment are the log files, can anyone help how to fix this problem and why it happends?
Thanks
Error.zip
pentaho.log
Attached Files

Calculate the difference between two identical strems

$
0
0
Hi,

I have two identical strems and I would like to get the difference between these two streams. Below is the statement that I use in the database, however I would like to do this before inserting the data into the database.

SELECT
a.Col1,
a.Col2,
a.Col3,
a.SalesEur - b.SalesEur
FROM dbo.tableA a
LEFT JOIN dbo.TableB b ON( a.Col1 = b.Col1 AND a.Col2 = b.Col2 AND a.Col3 = b.Col3)

Cannot Open any Pentaho Jobs

$
0
0
I have just installed 5 new licenses and I cannot open any of my Pentaho Jobs:

org.pentaho.di.core.exception.KettleXMLException:
Unable to load job info from XML node
There was an error while reading the shared objects (continuing load) : {0}
Unexpected problem reading shared objects from XML file : null
Error reading information from input stream
XML document structures must start and end within the same entity.



at org.pentaho.di.job.JobMeta.loadXML(JobMeta.java:1173)
at org.pentaho.di.ui.spoon.JobFileListener.open(JobFileListener.java:43)
at org.pentaho.di.ui.spoon.Spoon.openFile(Spoon.java:4550)
at org.pentaho.di.ui.spoon.Spoon.openFile(Spoon.java:4222)
at org.pentaho.di.ui.spoon.Spoon.openFile(Spoon.java:4159)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.pentaho.ui.xul.impl.AbstractXulDomContainer.invoke(AbstractXulDomContainer.java:313)
at org.pentaho.ui.xul.impl.AbstractXulComponent.invoke(AbstractXulComponent.java:157)
at org.pentaho.ui.xul.impl.AbstractXulComponent.invoke(AbstractXulComponent.java:141)
at org.pentaho.ui.xul.swt.tags.SwtToolbarbutton.access$100(SwtToolbarbutton.java:48)
at org.pentaho.ui.xul.swt.tags.SwtToolbarbutton$1.widgetSelected(SwtToolbarbutton.java:96)
at org.eclipse.swt.widgets.TypedListener.handleEvent(Unknown Source)
at org.eclipse.swt.widgets.EventTable.sendEvent(Unknown Source)
at org.eclipse.swt.widgets.Widget.sendEvent(Unknown Source)
at org.eclipse.swt.widgets.Display.runDeferredEvents(Unknown Source)
at org.eclipse.swt.widgets.Display.readAndDispatch(Unknown Source)
at org.pentaho.di.ui.spoon.Spoon.readAndDispatch(Spoon.java:1319)
at org.pentaho.di.ui.spoon.Spoon.waitForDispose(Spoon.java:7939)
at org.pentaho.di.ui.spoon.Spoon.start(Spoon.java:9190)
at org.pentaho.di.ui.spoon.Spoon.main(Spoon.java:654)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.pentaho.commons.launcher.Launcher.main(Launcher.java:92)
Caused by: org.pentaho.di.core.exception.KettleXMLException:
There was an error while reading the shared objects (continuing load) : {0}
Unexpected problem reading shared objects from XML file : null
Error reading information from input stream
XML document structures must start and end within the same entity.


at org.pentaho.di.job.JobMeta.loadXML(JobMeta.java:1000)
... 27 more
Caused by: org.pentaho.di.core.exception.KettleXMLException:
Unexpected problem reading shared objects from XML file : null
Error reading information from input stream
XML document structures must start and end within the same entity.

at org.pentaho.di.shared.SharedObjects.<init>(SharedObjects.java:165)
at org.pentaho.di.job.JobMeta.readSharedObjects(JobMeta.java:1192)
at org.pentaho.di.job.JobMeta.loadXML(JobMeta.java:995)
... 27 more
Caused by: org.pentaho.di.core.exception.KettleXMLException:
Error reading information from input stream
XML document structures must start and end within the same entity.
at org.pentaho.di.core.xml.XMLHandler.loadXMLFile(XMLHandler.java:637)
at org.pentaho.di.core.xml.XMLHandler.loadXMLFile(XMLHandler.java:558)
at org.pentaho.di.core.xml.XMLHandler.loadXMLFile(XMLHandler.java:539)
at org.pentaho.di.shared.SharedObjects.<init>(SharedObjects.java:108)
... 29 more
Caused by: org.xml.sax.SAXParseException; lineNumber: 170; columnNumber: 1; XML document structures must start and end within the same entity.
at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
at javax.xml.parsers.DocumentBuilder.parse(Unknown Source)
at org.pentaho.di.core.xml.XMLHandler.loadXMLFile(XMLHandler.java:614)
... 32 more


at org.pentaho.di.job.JobMeta.loadXML(JobMeta.java:1000)
... 27 more

Who am I? (PDI processes and Active Directory)

$
0
0
Hi all,

PDI Noob here with a question...

Short Version: If a user runs a transformation on a Windows server, can it be made to run under that user's Active Directory account? (Please say "yes"). If not, how can a transformation tell who launched it?

Long Version: Real Soon Now we will have a shiny new PDI server on Windows integrated with Active Directory.
But I am not yet sure what "integrated" will turn out to mean. I am hoping it will mean:


  • Nobody has to explicitly provide any credentials when running a transformation on the server (or have them baked into a .ktr file / set in an environment variable etc);
  • The transformations will run under my account not some system account;
  • If my transformation connects to SQL Server using Integrated Security, my credentials will be used;


Is that how things work?
If it isn't, is there at least a way a transformation can tell who launched it so this information can be passed to a service?

Ed

How To Interpret Merit Scores With WrapperSubset

Execute SQL script

$
0
0
database: mysql
error : create produce
why ?
1.jpg
code:



DELIMITER //


CREATE PROCEDURE proc7(IN s INT)


begin
declare i int;
declare s int;
set s = 0;
set i = 1;
while i<=n do
set s =s+i;
set i=i+1;
end while;
select s;
end;
//


DELIMITER ;

error :

2016/01/06 15:33:20 - test1.0 - ERROR (version 6.0.0.0-353, build 1 from 2015-10-07 13.27.43 by buildguy) : An error occurred, processing will be stopped:
2016/01/06 15:33:20 - test1.0 - Couldn't execute SQL: DELIMITER //
2016/01/06 15:33:20 - test1.0 -
2016/01/06 15:33:20 - test1.0 - CREATE PROCEDURE proc7(IN s INT)
2016/01/06 15:33:20 - test1.0 -
2016/01/06 15:33:20 - test1.0 - begin
2016/01/06 15:33:20 - test1.0 - declare i int
2016/01/06 15:33:20 - test1.0 -
2016/01/06 15:33:20 - test1.0 - You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near 'DELIMITER //
2016/01/06 15:33:20 - test1.0 -
2016/01/06 15:33:20 - test1.0 - CREATE PROCEDURE proc7(IN s INT)
2016/01/06 15:33:20 - test1.0 -
2016/01/06 15:33:20 - test1.0 - begin
2016/01/06 15:33:20 - test1.0 - declare i int' at line 1
2016/01/06 15:33:20 - test1.0 - ERROR (version 6.0.0.0-353, build 1 from 2015-10-07 13.27.43 by buildguy) : 错误初始化步骤[test1]
2016/01/06 15:33:20 - test - ERROR (version 6.0.0.0-353, build 1 from 2015-10-07 13.27.43 by buildguy) : 步骤 [test1.0] 初始化失败!
2016/01/06 15:33:20 - test1.0 - Finished reading query, closing connection.
2016/01/06 15:33:20 - Spoon - ERROR (version 6.0.0.0-353, build 1 from 2015-10-07 13.27.43 by buildguy) : test: preparing transformation execution failed
2016/01/06 15:33:20 - Spoon - ERROR (version 6.0.0.0-353, build 1 from 2015-10-07 13.27.43 by buildguy) : org.pentaho.di.core.exception.KettleException:
2016/01/06 15:33:20 - Spoon - 无法初始化至少一个步骤. 执行无法开始!
2016/01/06 15:33:20 - Spoon -
2016/01/06 15:33:20 - Spoon -
2016/01/06 15:33:20 - Spoon - at org.pentaho.di.trans.Trans.prepareExecution(Trans.java:1155)
2016/01/06 15:33:20 - Spoon - at org.pentaho.di.ui.spoon.trans.TransGraph$28.run(TransGraph.java:3991)
2016/01/06 15:33:20 - Spoon - at java.lang.Thread.run(Thread.java:744)
2016/01/06 15:33:20 - test - ERROR (version 6.0.0.0-353, build 1 from 2015-10-07 13.27.43 by buildguy) : 错误被检测到!
2016/01/06 15:33:20 - test - ERROR (version 6.0.0.0-353, build 1 from 2015-10-07 13.27.43 by buildguy) : 错误被检测到!
Attached Images

Is it possible to execute a job which repeats itself from JAVA?

$
0
0
I want to execute a job which repeats itself every several minutes and I want to execute the Job from JAVA.

However, I do not want to hold the thread since I have other jobs to be executed too.

Any ideas? Thanks~

Pivot Subtotal BtableQ

$
0
0
Hi All,
I have date dimension with level : Year, Quarter, Month

I put [Date.Months].[Year] , [Date.Months].[Quarter], [Date.Months].[Month] in pivot dimension, i have check the pivot subtotal and grandtotal.
And The Result is : 2015 Q1 Jan, 2015 Q1 Feb, 2015 Q2 Mar , 2015 Q2 Apr, etc.
I want the subtotal for the Q1 , Q2, etc. How to do it ?
I want like this : 2015 Q1 Jan, 2015 Q1 Feb, 2015 Q2 Mar , Subtotal Q1, 2015 Q2 Apr, etc.

Any tutorials on how to implement Schema Workbench Calculated Member / MDX formula?

$
0
0
I have a mondrian schema, with standard measures.
I'm new to MDX so I guess this makes my task a bit more difficult, im struggling with the syntax of the mdx formula.

Is there any way to test my mdx formula to see the results it returns?

Does anyone have any links to some information / tutorials / guides on implementing a calculated member in Schema Workbench to help me?

Thanks
Jason

Display the logs of a job running inside another job

$
0
0
Hi. I am running a job which calls another job. But, the output just logs details about the main job.
How is it possible to obtain the log of the inner job ? I tried to use the "write log" step, but it doesn't work either.Thanks for help.

Keycloak SSO Integration

$
0
0
hi to all,
i'm early new with Pentaho.
That's my use case.
I use SSO Jboss Keycloak as my single sign on server. I have already developed few applications which already use it, now this is the time to wrap Pentaho BI inside the SSO mechanism.

I have read Pentaho authentication is based on Spring Security and luckily Keycloak offers a Spring Security Adapter by which i can secure a spring secured application with keycloak authentication.
Can someone help me by listing some steps i have to do in order to develop an authentication keycloak provider to be integrated inside pentaho authentication process?
I mean that:
- when i try to login pentaho web application i will be redirected to keycloak SSO login form.
- when i try to access pentaho web applications, being already authenticated by other applications, i can do it without any credential request.

Coming more into detail, i have successfully tried to develop a spring secured application using keycloak SSO.
It could simply done by modifying the security-context.xml as the following listing:

<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:context="http://www.springframework.org/schema/context"
xmlns:security="http://www.springframework.org/schema/security"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans
http://www.springframework.org/schem...ring-beans.xsd
http://www.springframework.org/schema/context
http://www.springframework.org/schem...ng-context.xsd
http://www.springframework.org/schema/security
http://www.springframework.org/schema/security/spring-security.xsd">


<context:component-scan base-package="org.keycloak.adapters.springsecurity" />


<security:authentication-manager alias="authenticationManager">
<security:authentication-provider ref="keycloakAuthenticationProvider" />
</security:authentication-manager>


<bean id="adapterDeploymentContextBean" class="org.keycloak.adapters.springsecurity.AdapterDeploymentContextBean" >
<constructor-arg value="/WEB-INF/keycloak.json" />
</bean>
<bean id="keycloakAuthenticationEntryPoint" class="org.keycloak.adapters.springsecurity.authentication.KeycloakAuthenticationEntryPoint" />
<bean id="keycloakAuthenticationProvider" class="org.keycloak.adapters.springsecurity.authentication.KeycloakAuthenticationProvider" />
<bean id="keycloakPreAuthActionsFilter" class="org.keycloak.adapters.springsecurity.filter.KeycloakPreAuthActionsFilter" />
<bean id="keycloakAuthenticationProcessingFilter" class="org.keycloak.adapters.springsecurity.filter.KeycloakAuthenticationProcessingFilter">
<constructor-arg name="authenticationManager" ref="authenticationManager" />
</bean>


<bean id="keycloakLogoutHandler" class="org.keycloak.adapters.springsecurity.authentication.KeycloakLogoutHandler">
<constructor-arg ref="adapterDeploymentContextBean" />
</bean>


<bean id="logoutFilter" class="org.springframework.security.web.authentication.logout.LogoutFilter">
<constructor-arg name="logoutSuccessUrl" value="/" />
<constructor-arg name="handlers">
<list>
<ref bean="keycloakLogoutHandler" />
<bean class="org.springframework.security.web.authentication.logout.SecurityContextLogoutHandler" />
</list>
</constructor-arg>
<property name="logoutRequestMatcher">
<bean class="org.springframework.security.web.util.matcher.AntPathRequestMatcher">
<constructor-arg name="pattern" value="/sso/logout**" />
<constructor-arg name="httpMethod" value="GET" />
</bean>
</property>
</bean>
<security:http auto-config="false" entry-point-ref="keycloakAuthenticationEntryPoint" use-expressions="true">
<security:custom-filter ref="keycloakPreAuthActionsFilter" before="LOGOUT_FILTER" />
<security:custom-filter ref="keycloakAuthenticationProcessingFilter" before="FORM_LOGIN_FILTER" />
<security:intercept-url pattern="/admin/*" access="hasRole('MYROLE')" />
<security:custom-filter ref="logoutFilter" position="LOGOUT_FILTER" />
</security:http>
</beans>


So i m wondering if this mechanism could be easily imported into petaho application-security-context.xml

Thanks a lot

Prepare DEvelopment Plataform

$
0
0
Hi.

Happy new year for everybody.

Can someone post the complete steps to configure eclipse to build pentaho 6? I've been trying for a week but something keeps not working.

I used tutorial for 5 as base.

Thanks in advance.

JP

Carte execution ID vs JOB execution log

$
0
0
Hi,
I've developed a job flow with lots of transformations and I also enabled job and transformation logging.
I've enabled carte and I can execute the job by invoking a URL.
The return after URL call is:
<webresult>
<result>OK</result>
<message>Job started</message>
<id>549387d0-e621-4b6f-aa63-2783a2116e7c</id>

</webresult>

What I want to do is to check at log tables from jobs and transformations wich carte execution ID (in this case "549387d0-e621-4b6f-aa63-2783a2116e7c") generated the log lines.
I can see lots of possible columns like CHANNEL_ID, PARENT_CHANNEL_ID and ROOT_CHANNEL_ID but none seems to save the CARTE execution id.

Does anyone know how we can check wich carte execution ID generated the job/transformation log lines.

Many thanks in advance,
Ronaldo Akamine

Incremental Load best practices......

$
0
0
Hi Everyone,

I amtrying do incremental load. And it's type 1 dimension. The old data getsreplaced new changes.

I amgetting the max(last_modified_ts) from the target table using "Table Input1" and pass it thro' another "Table Input 2" in where clause to
get onlythe changed records from the source. And up to this part, it works.

Afterthat, I need to compare the column "col_pk" from "Table Input2" result set against "col_pk" of "Target table".
If
TableInput 2.col_pk = target table.col_pk then compare the remaining columns in thetable to check if there was any change. If so, update that record.
If
TableInput 2.col_pk = target table.col_pk and when compared there is no changes toremaining columns then don't update.
If
TableInput 2.col_pk not found in target table.col_pk then insert into target table.

It's kindof merge. So, I tried merge rows - diff. The flagfield for insert and updatecomes as "new" and the not changed records comes as"deleted".
Or I can"delete" the new records from target and then "insert". Inboth cases, I am not sure, how to pass this "flagfield" to triggerthe delete and insert or
updateand insert. Please find attached document that shows merge rows-diff setup andthe results set.

Pleasehelp.

Thanks,
Raji.
Attached Files

PDI and implementing Machine Learning (ML) algorithm

$
0
0
Searching for ideas, resources, documentation, demos etc...

Requirement is to leverage PDI (Enterprise Edition) capability to execute the following work steps.

1) Extract/Filter a data set (thousands of rows, csv file)
2) Pass the ENTIRE data set to a ML algorithm (writing in Python) which will add additional columns to every row.
3) Capture the ENTIRE data set from step #2 and insert to a RDBMS.

Looking for options/insight on how to implement Step #2.

I understand there is a R Language plug in. Is there one for Python ML?

"How do I print CTools dashboards in Pentaho?"

$
0
0
How do I print CTools dashboards in Pentaho?
This is a question that often keeps coming back. Even for my own reference I decided to just compile the answer that I always give in a blog post links to the relevant resources

This is a simple question with a not so simple answer. Screen and paper obey such different rules that really depends on what the customer is looking for. There are 3 common approaches:

1. The browser button

Just use the browser print button (or a button in the dashboard that uses window.print() as the code). This has the obvious advantages (dead easy) and the obvious advantages as well (the actual output depends on the browser)
2. PhantomJS on server side



Use phantomjs on the server side. Not drastically different than the previous but possible to control the output. I know of a customer that even implemented scheduling of PDFs this way. Harder to setup. Some links:

My favorite approach - PRD



The truth is that screen and paper is not the same. There's no 1:1 mapping, even with CSS media queries the result will be, at best... meh.


My personal favorite is actually leveraging the most appropriate tool on our stack to print: PRD. My recommendation is that we build a report using the resources of the dashboard (queries and charts) and build the pixel perfect representation of the dashboard as a report. References:



Have fun!


-pedro

More...
Viewing all 16689 articles
Browse latest View live


Latest Images

<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>