Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

R script executor crashes for big models

$
0
0
Hi, I am trying to use R script executor using Pentaho EE 7.0.0.0-25 (not free Execute R script) in my ETL. I already have prepared model from R and now I just want to apply it for new data.
The model is read using this statement
Code:

fit <-
  readRDS(file = qq("c:\\path\\to\\model.rds"))

but the model is quite soficticated and therefore has over 400MB. So when I press "Test script" (or launch the transformation) after few second whole Pentaho crashes with classic Java(TM) Platform SE binary has stopped working.
The problem is most probably in size of the model because when I switch it to more simple one (with 1MB) it works.

My first hint was not enough memory so I tried it on another machine but memory usage didn't go over 8,2GB when PDI crashed - and yes, I have change the Xmx parameter in spoon.bat to 32048MB.

Running the script straight in R works as well, so my second hint is "isn't there any other memory limit for plugins?"

And for the record the plugin was installed using Marketplace so no dlls copying. Environment variables are probably set up correct if the script works for small model.

Does anyone have any idea how to fix this?

Cube Publishing Error

$
0
0
I've been having problems viewing my cube. I'm using Pentaho BI Server 4.8.0-stable, Schema Workbench 3.2.2-stable, jdk 1.7.0_79, on Windows 10.

My desired result is a cube with one fact dimension, and three additional dimensions. One of these dimension being a Time dimension, including month and year. My cube publishes fine, viewing it on Pentaho is a problem.


The xml code:
<Schema name="Recall">
<Cube name="Cubo Recalls" cache="true" enabled="true">
<Table name="fato_campanha_recall" schema="public" alias="Fato Campanha Recall">
</Table>
<Dimension type="StandardDimension" foreignKey="empresa_chave" highCardinality="false" name="Empresa">
<Hierarchy name="Empresa" hasAll="true">
<Table name="empresa" schema="public" alias="Empresa">
</Table>
<Level name="Fornecedor" column="fornecedor" type="String" uniqueMembers="false" levelType="Regular" hideMemberIf="Never">
</Level>
</Hierarchy>
</Dimension>
<Dimension type="StandardDimension" foreignKey="produto_recall_chave" highCardinality="false" name="Produto_Recall">
<Hierarchy name="Produto_Recall" hasAll="true">
<Table name="produto_recall" schema="public" alias="Produto_Recall">
</Table>
<Level name="Assunto" column="assunto" type="String" uniqueMembers="false" levelType="Regular" hideMemberIf="Never">
</Level>
<Level name="Classificacao" column="classificacao" type="String" uniqueMembers="false" levelType="Regular" hideMemberIf="Never">
</Level>
<Level name="Objeto" column="objeto" type="String" uniqueMembers="false" levelType="Regular" hideMemberIf="Never">
</Level>
<Level name="Defeito" column="defeito" type="String" uniqueMembers="false" levelType="Regular" hideMemberIf="Never">
</Level>
</Hierarchy>
</Dimension>
<Dimension type="TimeDimension" foreignKey="tempo_chave" highCardinality="false" name="Tempo">
<Hierarchy name="Tempo" hasAll="true">
<Table name="tempo" schema="public" alias="Tempo">
</Table>
<Level name="Mes" column="mes" ordinalColumn="mes" type="Numeric" uniqueMembers="false" levelType="TimeMonths" hideMemberIf="Never">
</Level>
<Level name="Ano" column="ano" ordinalColumn="ano" type="Numeric" uniqueMembers="false" levelType="TimeYears" hideMemberIf="Never">
</Level>
</Hierarchy>
</Dimension>
<Measure name="Quantidade_Afetados" column="quantidade_afetados" datatype="Integer" formatString="#,###" aggregator="sum" visible="true">
</Measure>
<Measure name="Quantidade_Atendidos" column="quantidade_atendidos" datatype="Numeric" formatString="#,###" aggregator="sum" visible="true">
</Measure>
</Cube>
</Schema>




Pentaho.log:

2017-01-02 15:46:38,381 WARN [org.pentaho.hadoop.shim.HadoopConfigurationLocator] Unable to load Hadoop Configuration from "file:///C:/pentaho/biserver-ce/pentaho-solutions/system/kettle/plugins/pentaho-big-data-plugin/hadoop-configurations/mapr". For more information enable debug logging.
2017-01-02 15:46:41,329 WARN [org.pentaho.reporting.libraries.base.boot.PackageManager] Unresolved dependency for package: org.pentaho.reporting.engine.classic.extensions.datasources.cda.CdaModule
2017-01-02 15:46:41,460 WARN [org.pentaho.reporting.libraries.base.boot.PackageSorter] A dependent module was not found in the list of known modules.
2017-01-02 15:46:59,917 WARN [org.apache.commons.httpclient.HttpMethodBase] Going to buffer response body of large or unknown size. Using getResponseBodyAsStream instead is recommended.
2017-01-02 15:47:03,511 WARN [org.apache.axis2.description.java2wsdl.DefaultSchemaGenerator] We don't support method overloading. Ignoring [public java.lang.String serializeModels(org.pentaho.metadata.model.Domain,java.lang.String,boolean) throws java.lang.Exception]
2017-01-02 15:50:50,925 WARN [org.pentaho.platform.engine.services.runtime.TemplateUtil] not found: [Measures].[Quantidade_Afetados]



Any and all suggestions are welcome. I'm new to Pentaho so please be as detailed as possible :)
Attached Images

"Put a file with SFTP" failed

Connection Termination for Long Running Query

$
0
0
Hi - We are using the PDI CE 6.1 and trying to load tables into Redshift. While processing high volume data, it appears that pentaho lost connection to the Redshift database and I don't see any processes running on the EC2 instance (EC2 instance is the Pentaho's launching platform). Whereas, the query launched by the process was still running on the Redshift cluster visible through AWS console. Normally the Pentaho log file receives heart beat signal upto 2 hours or so. But it stops after that and all subsequent processes in the same pentaho job file are not executed further.

I tried to use the PDI CE 7.0 to run the jobs (however jobs were created using 6.1) but faced the same issue.

Do anyone of you faced such issue? Is there any solution to resolve this issue?


Thanks,
AG

how to count number of header columns in text file input

$
0
0
Hi Team,

I have reading data from text file input.how to count number of header columns in text file input.

How to connect PDI with MS SQL Server 2016

$
0
0
Step1: MS SQL Server
ServerName: MINK-NB
Authen username: sa password: mink

Step2: PDI
2.1 I download sqljdbc_6.0 from microsoft and copy sqljdbc.jar / sqljdbc4.jar / sqljdbc41.jar / sqljdbc42.jar to path pdi-ce-7.0.0.0-25\data-integration\lib already.
2.2 I create connection in PDI.
Connection Type: MS SQL Server Access: Native(JDBC)
Hostname: localhost << I'm not sure
Database name: MyDB
Instance name: MINK-NB (ServerName from Step1)
Username: sa
Password: mink

But I can't working that it show .... How can I do?

Error connecting to database [con] :org.pentaho.di.core.exception.KettleDatabaseException:
Error occurred while trying to connect to the database

Driver class 'net.sourceforge.jtds.jdbc.Driver' could not be found, make sure the 'MS SQL Server' driver (jar file) is installed.
net.sourceforge.jtds.jdbc.Driver


org.pentaho.di.core.exception.KettleDatabaseException:
Error occurred while trying to connect to the database

Driver class 'net.sourceforge.jtds.jdbc.Driver' could not be found, make sure the 'MS SQL Server' driver (jar file) is installed.
net.sourceforge.jtds.jdbc.Driver

dynamic multiselect filters in pentaho dashboard

$
0
0
Hi All,

I need to implement dynamic filters in my dashboard. consider there are 5 multi select filters if i select one then all others should refresh as per the selected value and the selected filter should not refresh again if i select another filter.
can i know how/ in which order listeners should be used to achieve this in multi select component. please suggest me.

thank u in advance.

Problem with files' naming - Pentaho User Console 6.1

$
0
0
Hi, I have an issue. When I upload a new file from Pentaho Report Designer I put Name and so on....

But instead of showing the nema, the User Console show me the name of the .prpt file.

Rename is the only thing I can do or there is some bug or option?

Thanks

Where are users and roles stocked in pentaho

$
0
0
Hi,
I'm using pentaho 7 EE and trying .
I didn't found where are users/roles and the mapping between roles/permissions stored?

ANy help please?

Problem connecting to HBASE table.

$
0
0
Hi All,

I am using pdi-ce-6.1.0.1-196 version to connect Hbase table through HBase Input component.

Hbase version is 1.1.2 and Zookeeper version is 3.5.1

I have changed all necessary jars in location: D:\pdi-ce-6.1.0.1-196\data-integration\plugins\pentaho-big-data-plugin\hadoop-configurations\hdp23

But I am still unable to Get Mapped tables in Hbase Input component. When trying to load mapped tables I am getting the below exception. Please help.

java.io.IOException: java.lang.IllegalArgumentException: Not a host:port pair: PBUF

I can connect to hadoop cluster and tetsted connection is successful.

Java exception while running Spoon.bat

$
0
0
Dear All,

I am facing issues "A Java Exception Has Occured" while trying to Spoon.bat from pdi-ce-7.0.0.0-25 in Windows 7, 64 bit. The JDK version used is 1.7.0_79.
After reading somewhere I had changed javaw to java in spoon.bat. However, after that I am not getting any error but spoon.bat is also not running. There's one cmd window popping up and closing immediately.

Can anyone of you please let me know, what is going wrong, and how to resolve ? Also for people who are succesfully using this version of Spoon, can you please let me know your Java version and Windows version, and also if any additional settings had to be done.

Thanks in advance for your help.
Tanisha

json input with different fields (meta data injection)

$
0
0

I have several grouped parameters.

Example:
1.Group:
table
: channel
field
: via

2.Group:
table
: ticket_tags
field
: tags

And a following transformation. Till now I was able to execute everything only for the first group. But I want to generalize it, so that I read the file with the grouped parameters and the transformation executes for each set of the parameters. I assume I need meta data injection step. However, I am not sure how I should change the Fields/Path in json input to depict the variable from meta data injection, namely hub_field, for instance. For now I have there : $.ticket_events[].${hub_field}* , but it shows me an error:

2017/01/0315:28:53- JSON Input.0-We can not find any data with path [$.ticket_events[*].${hub_field}]!

The structure of the json file is:
Code:

{
 
"end_time":1481672949,
 
"ticket_events":[

   
{
     
"id":110831272025,
     
"via":"Web Widget",
     
"child_events":[
       
{
         
"event_type":"Create",
         
"id":110831272085,
         
"tags":[
           
"web_widget"
         
]
       
}
       
]
   
}
   
]
}


So, I want the via field from ticket_events and tags field from ticket_events.child_events BUT viaand tags are comping from metadata injection via hub_field.


And I have another transformation with only JSON Input which reads the json grouped parameters from above. The output is following:

Reading emails and saving in proper HTML format

$
0
0
Hi guys,

I am trying to read emails and save email in HTML format using Job's Get mails (POP3/IMAP) task. As shows in below screenshot, I have given target filename pattern as saving in HTML format.
Get emails from POP - Target settings.jpg
But I see if I open the saved html file, it's not in proper HTML format. My email is having Image pasted in Body part, as shown in below screen.
Sample Email.jpg

If I open saved HTML file in browser, it doesn't open.

Any pointers, How do I save Email in Proper HTML format?
Thanks,
Nilesh
Attached Images

Measure Has values But Drilling Down Returns Nothing

Error importing and exporting the BI-repo

$
0
0
When I am trying to import or export the bi-server repository, I am getting this error message when exporting: CommandLineProcessor.ERROR_0002 - Invalid ClientResponse received in performREST()

I have restarted the JVM but still it is the same. Does anybody know about this issue? I am using Pentaho 5.4.0.8.

Thanks!

CCC Bar chart : displaying a threshold

$
0
0
Hi,

I have a bar chart displaying a monthly ratio (e.g. amount of sales in a country ) : I have the month on horizontal axis and the height of each bar is the amount.

Let's assume that the boss sayed 'the monthly amount of sales must be at least 10000€' : I would like to draw an horizontal line with y-coordinate = 10000

How can I do that?

Thank you

Benoit

showing trend graph in Table Component

$
0
0
Hi

Is their is any way to display trend graph in Table Component's Coloumn?

Matrix dashboard

$
0
0
Hello,
I tried to create a reporting I want to have an array matrix board as the example in the file attachment.
I used the analysis crossed but I'd be stuck in the stage of total and I can not move maleureusement, thank you for helping me
Matrice.jpg
Attached Images

Use connection parameter in MDX

Problem with plug-in creation : Mvc and Xul management

$
0
0
Hello there,

I am a young Java developper and i am currently trying to create a new plug-in for PDI. I want to implement an HTTP Client to Spoon, allowing the user to access my own transformation, jobs, features, ect...

My problem is i can't reach to change some Xul components from my Java code.

As an overlay, i put a new menupopup to the main menu bar. Some of the items inside possess a "disabled" tag :

Code:

<menuitem id="open-item" label="Ouvrir..." disabled="true"/>
I want to change the disabled value to "false" when i'm connected. So i tried the following :

- To use some XulComponent objects, and i managed to change this kind of object's disabled value, but it doesn't reflect on Spoon. I suppose it's because the change happen in the object only.

Code:

XulComponent xc =  (XulComponent) document.getElementById("myMenu").getElementById("open-item");
xc.setAttribute("disabled", "false");

- To use Bindings, but as the bind() method is deprecated, i can't use the tutorial method from here : http://wiki.pentaho.com/display/Serv...L+Applications

So i turned to some BindingFactory as it seems to be the new way to make it, but when i change my own object parameters value after a bind, it doesn't seems to change the value of XulComponent or the value in Xul file either.

Here is the binding code (inside my handler object) :

Code:

public void test(){

    XulComponent xc =  (XulComponent) document.getElementById("myMenu").getElementById("open-item");       
    BindingFactory bf = new DefaultBindingFactory();
    bf.setDocument(document);
    Binding bind = bf.createBinding(mm1, "okEnabled", xc, "disabled"); // mm1 is just an object with a name(string) and a enabled(boolean) parameters
    mm1.setOkEnabled(false);

    }

and the xul file :
Code:

<?xml version="1.0" encoding="UTF-8" ?>
<?xml-stylesheet href="chrome://global/skin/" type="text/css"?>
<overlay id="my-overlay" title="my overlay"
    xmlns="http://www.mozilla.org/keymaster/gatekeeper/there.is.only.xul"
    height="600"
    width="800" xmlns:pen="http://www.pentaho.org/2008/xul">


    <menubar id="spoon-menubar">
        <menu id="myMenu" label="my Menu">
            <menupopup id="myMenu-popup">
                <menuitem id="connect" label="Se connecter"  command="myMenu.showLogin()"/>
                <menuseparator id="myMenu-separator-1"/>
                <menuitem id="open-item" label="Ouvrir..." disabled="true"/>
                <menuitem id="save-as-item" label="Enregistrer sous..." disabled="true" />
                <menuseparator id="myMenu-separator-2"/>
                <menu id="publish" label="Publier" >
                    <menupopup id="publish-popup">
                        <menuitem id="publish-transformation" label="Transformation" disabled="true"/>
                        <menuitem id="publish-job" label="Job" disabled="true"/>
                    </menupopup>
                </menu>
                <menuseparator id="myMenu-separator-3"/>
                <menuitem id="close-item" label="Fermer" command ="myMenu.showLoginAlert()"/>
            </menupopup>
        </menu>
    </menubar>   
</overlay>

As i'm still inexperienced with Pentaho, idon't know if i am on the good way to solve it. Or am i totally wrong? is there another way to make it properly?

Thank you in advance for your answers, and sorry if i made some mistake with my english. If you need any others informations, please just tell me and i'll try to answer as fast as possible.
Viewing all 16689 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>