Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

Carte service is down

$
0
0
Hi,
I have a few questions regards to Pentaho configurations:
1. I see there are -Xms, -Xmx and -MaxPermSize in Spoon.bat/Spoon.sh. I understand that Xms and Xmx is minimum and maximum memory allocation to Spoon. But I am not sure what MaxPermSize is and what it doing. We have set the xms and MaxPermSize to 8GB and xmx to 32GB. Does this help to boost Spoon's performance? We have 256GB in the server and we split to 4 Spoon services, allocate 32GB each.


2. Carte service is down when Pentaho process large files (10-12million records text files) without any indication and we would need to restart the carte service. How can we improve this? Is Carte responsible to invoke Spoon to run transformations?


3. How can we store the details of the log file into database?


4. The service crashed after processed a large file and not able to process the next file, and we needed to restart the services again. I believe that is because the memory is not fully being release after processed the large file. How can we flush/release the machine memory if the memory has hit the maximum after processed the large file?


5. How can we improve the Spoon performance if the file is too large (10-12million records text files). The duration taken now is around 4 hours. How can we shorten by 40-50% of the duration?


Pentaho exact version: 4.0.1


Appreciate your help on my queries.
Best Regards,
Kar Ho

Filter Component - search on server side

$
0
0
I have a Filter Component with a dataset containing 2,000 rows. I have a Page Size of 300. I want to ensure that if a user enters a search term, that the search occurs on the server rather than in the displayed items. Currently when a user searches, it only searches the visible items, from the first page (300). I want the search enable users to find the item they are looking for without having to scroll through many pages.

I did have this working at some time in the past, I believe by adding the following code to the Advanced Settings in the filter component

Code:

function f(){
    return {
        component: {
            search: {
                serverSide: true
            }
        },
    };
}

But it's definitely not working now. Maybe I'm wrong and it never worked, or maybe an update to CDE has made it not work?

I'm running Pentaho 7.0 / CDE 7.0.0.1-37 (7.0.0.1)

Connect to Google Bigquery with generic connection

$
0
0
I want to connect to bigquery with a generic database connection.

I've managed to do that with the old driver 'bqjdbc-1.4-standalone.jar' but it doesn't support Standard SQL and using the Legacy SQL is very limiting.

I was trying to use the new drivers from simba and copied all the jar files to lib folder https://cloud.google.com/bigquery/pa...simba-drivers/

The connection string URL was: jdbc:bigquery://https://www.googleapis.com/bigquery/v2:443;ProjectId=[project id];OAuthType=0;OAuthServiceAcctEmail=[name]@[project label].iam.gserviceaccount.com;OAuthPvtKeyPath=[path to p12 key file]
Driver Class Name: com.simba.googlebigquery.jdbc42.Driver

I receive the following error:
error connecting to database: (using class com.simba.googlebigquery.jdbc42.Driver)
com.google.api.client.http.javanet.NetHttpTransport.<init>(Lcom/google/api/client/http/javanet/ConnectionFactory;Ljavax/net/ssl/SSLSocketFactory;Ljavax/net/ssl/HostnameVerifier

Any ideas?

Thank you!

Linux batch file log error

$
0
0
Hi Sir,Madam,

We are using below Linux script to generate the log files for ETL running activity. We are facing issues with log file extension, there Is a special character is placing at end.
Sorry my bad luck, I am not getting attachment option to add my sample snap shot. Could you please correct me if below script is wrong

ETL_JOB_20180726_1148.log ( ETL_JOB_20180726_1148(special character).log

echo "Linux_batch_file_etl"

cd /data-integration

NOW=$(date +"%Y%m%d_%H%M")
sh kitchen.sh -rep:"product" -job:"ROOT" -dir:/ -user:admin -pass:admin -logfile="/home/ubuntu/FILES/LOGS/ETL_JOB_$NOW.log" -level= Basic

f='logfile'; mv 'logfile' ${f//[^A-za-z0-9._-]/_}

Pentaho Community Meeting - PCM18! Bologna, Italy, November 23-25!

$
0
0


PCM 18!!

If you've been in one, no more words are needed, just go ahead and register! If you don't know what I'm talking about, just go ahead and register as well!

It's the best example of what Pentaho - how part of Hitachi Vantara - is all about. A very passionate group of people that are absolutely world class at what they do and still know how to spend a good time!


PCM17 group photo


Now shamelessly copy-pasting the content from it-novum:

Pentaho Community Meeting 2018

Pentaho Community Meeting 2018 will take place in Bologna from November 23-25. It will be organized by Italia Pentaho User Group and by it-novum, the host of PCM17. As always, it will be a 3-days event full of presentations, networking and fun and we invite Pentaho users of every kind to participate!

For PCM18 we will meet in the beautiful city of Bologna. The guys of Italia User Group will take care of the venue and the program. With Virgilio Pierini as group representative we not only have a Pentaho enthusiast but also a native of Bologna guiding us to the beautiful corners of the hometown of Europe’s oldest university!

What is Pentaho Community Meeting?

Pentaho Community Meeting is an informal gathering for Pentaho users from around the world. We meet to discuss the latest and greatest in Pentaho products and exciting geek stuff (techie track) as well as best practices of Pentaho implementations and successful projects (business track). Read this summary of Pentaho Community Meeting 2017 to learn more.

PCM18 is open to everyone who does something with Pentaho (development, extensions, implementation) or plans to do data integration, analytics or big data with Pentaho. Several Pentaho folks – architects, designers, product managers – will share their latest developments with us.

The event is community-oriented and open-minded. There’s room for networking and exchanging ideas and experiences. Participants are free to break off into groups and work together.

Call for Papers

For sure, this is intended to be a community event - for the community and by the community. To register your proposal for the agenda, please use the contact form to send a brief description including your name and title in English until September 30th.

Agenda

The agenda will be updated continuously, so stay tuned for updates! All updates will be posted on twitter, too.

Friday, November 23 | Hackathon

We start the three-day PCM with a hackathon, snacks and drinks. After a 2-hour hackathon, a highly esteemed jury will award the most intelligent/awkward/funny hacks.

Saturday, November 24 | Conference Day

Still a lot to be determined! We're still receiving papers


  • Welcome speech | Stefan Müller and the org team
  • The future of Pentaho in Hitachi Vantara | Pedro Alves, Hitachi Vantara
  • What's new in PDI 9.0 | Jens Bleuel, Hitachi Vantara
  • Useful Kettle plugins | Matt Casters, Neo4j (and founder of Kettle)
  • IoT and AI: Why innovation is a societal imperative | Wael Elrifai, VP for Solution Engineering - Big Data, IOT & AI, Hitachi Vantara
  • Pentaho at CERN | Gabriele Thiede, CERN
  • Pentaho User Group Italia
  • SSBI (Self Service BI ) - Pentaho Plugin Update | Pranav Lakhani, SPEC INDIA
  • Scaling Pentaho Server with Kubernetes | Diethard Steiner
  • Capitalizing on Lambda & Kappa Architectures for IoT with Pentaho | Issam Hizaji, Lead Sales Engineer, Data Analytics & IoT | Emerging & Southern


After the lunch, everybody splits up to join the business or the techie track.

Sunday, November 25 | Social Event

Brunch, sightseeing and... let´s see!


----

Anyway, believe me, you want to go! GO REGISTER HERE!




More...

Every 5 rows of data - consolidate to 1 row?

$
0
0
I am looking to consolidate every 5 rows of data into 1 row. I added a "Add sequence" step, but which step will allow me to create a single row from every 5?

Thanks in advance!!!

KP

Question about string lookups in MS-SQL

$
0
0
I'm reading in a text file and doing a sql lookup on one of the fields that is defined as String.
I'm trimming the field and it typically has 8-20 characters and defining the length as 255 on the text file input.



The key I'm looking up is defined as varchar(255) on the data base.
Our DBA was reviewing the activity on the server and noted that in the query the lookup is coming as nvarchar(4000).

I'm using sqljdbc4.jar for the drive rand currently pdi-cd-7.0.0.0-25 - though I will be upgrading to 8 shortly.

Is there a way to limit the size of what the sql query will be using or have it use nvarchar(255)?

Thanks.

File/Open or Browse Repository not showing all files

$
0
0
Hi, this manifests with both Linux pdi 6.0.1.0-386 and with Windows 5,3.0.0-213.

using File.Open or browse Repo I can't see any recent files except files created from spoon and saved into the repo file system structure (file system repo synced with a master using git). If I copy KTRs or KJBs from one directory to another in the repo tree I don't see the newly created files - equally I don't see new files that were added to the local repo from a git pull. Very strange behavior. I have checked file and directory privs - all OK, Also I have show hidden files set.

Anyone seen anything like this? Thanks in advance if anyone has..

Issue with S3 File Output component in PDI 8.1 Solved

$
0
0
Hi All,

We tried upgrading PDI 6.1 to 8.1 and found that the jobs using S3 File Output component were throwing errors.

When we checked the component , we found that the component was missing the ACCESS_KEY and SECRET_ACCESS_KEY placeholders.

We tried various options and finally ended up with a working solution.

Steps followed were :

1)Set two environment variables :

export AWS_ACCESS_KEY_ID=<<ACCESS_KEY>>
export AWS_SECRET_ACCESS_KEY=<<SECRET_ACCESS_KEY>>

2)Change the path of S3 as below :

s3n://s3n/<<bucket_name>>/<<folder_name>>/<<file_name>>

Thanks,
Pradeep

Login token XXXX is invalid or has expired

$
0
0
Hi,

Trying to extract data from Birst system ( cloud based dashboard and reporting system) which exposes data via SOAP API.

I'm able to generate the Login token using "webservice lookup" step. When I pass the same token to another "webservice lookup" step , I get the below error message.

Login token XXXX is invalid or has expired

The same works fine in SOAPUI when I manually run these API.

Message from Birst:
User login token has a validity of 10 minutes, so one need to make at least probing request like getjob/space status to keep same session alive. Apart from that Birst User Sessions are tied to particular server & this information resides in memory, One has to do cookie management in their web services code so that load balancer will redirect subsequent request from user to the same server from where the login token was issued originally.

How do we do cookie management in pdi?

Appreciate any help.

Regards
Raghu

Possible to change date formats without changing the Schema?

$
0
0
Hi, I'm wondering if there is a way to change the format in which a date appears in Analyzer? We have a date hierarchy including year, quarter, month, and day but also have field for week. The way we want 'week' to be displayed is in the format 'mm-dd-yy'. We have this in the database as a column called week_monday_start_name and it is in this format.
Is it possible for our users (who do not have access to the schema or any development aspect of pentaho) to change the format while creating a report?

Anyone have scripts to create windows service for PDIEE 8X

$
0
0
Installed the archive build of PDIEE 8 on our Windows Server 2012 box. Does anyone have a script to run the tomcat as a service instead of using the start-pentaho.bat file?

thanks
dan

pentaho and elasticsearch

$
0
0
Hi - I just downloaded PDI 8.0 version and trying to connect to elasticsearch using the elasticsearch bulk insert step.

However I see :
in this path :

pentaho\data-integration\plugins\elasticsearch-bulk-insert-plugin\lib

only : elasticsearch-2.2.0.jar file . Does this mean pentaho only will support elasticsearch of 2.2 version only ?
I tied copying the elasticsearch latest version - which is6.3 the , but I am not able to connect to the server. D I needto copy any othe jar files as well?

Any thoughts for me ? Appreciate all your time and help.

attachments issue

$
0
0
hi

What I did mistake i am not getting in this PDI forum, i am not able to add any images or attachments. kindly help me on this how can i resolve it

To whom i need to request on this issue.

than you

Mondrian cannot use JDBC Data source which uses Oracle TNS name

$
0
0
Hello All,

Could someone please help with the following issue:

I have 2 JDBC Data sources created in Pentaho BA Server:

1. Oracle Connection with providing user credentials, host, port and SID as database name.
2. Oracle Connections with providing user credentials and TNS name as database name, since if there is SERVICE NAME then I should use TNS name.

Both above connections ping successfully.

When I use the #1 Option as JDBC Data source for my Cube schema - everything works.
But, when I use the #2 Option, cube cannot be loaded and the following exception occurs in the log file:

java.lang.RuntimeException: Unable to get connnection jdbc:mondrianataSource=Report DB; EnableXmla=false; overwrite=false; Locale=en_US; Catalog=mondrian:/My_Cube_Schema; UseContentChecksum=true
at com.pentaho.analyzer.service.impl.OlapConnectionManagerImpl.createConnection(SourceFile:152)
at com.pentaho.analyzer.service.impl.OlapConnectionManagerImpl.createConnection(SourceFile:75)
at com.pentaho.analyzer.service.impl.a.getConnection(SourceFile:55)

So, the question is, how could I achieve the connection for my cube, if I have just SERVICE NAME as for connection to DB.

Any help would be appreciated!

Thanks,
Bagdat

Cygwin with Pentaho

$
0
0
hi Sir , Madam,

we are working on Cygwin terminal with PDI 3.2 version. our etl assets are exist on Linux box and we are trying to connect ETL with Cygwin software. We are able to connect PDI successfully from my local machine but it is very very slow .

Could you please tell me know if anyone faced this type of problem, not sure where i need to post this problem if i faced like this. Please help me

Thank you

How do I drop a link with an iloader call?

$
0
0
Hi guys,
This may be more of an iLoader question but I'm using Kettle to call the iLoader script so I'll post here in case someone knows. I have a simple job that invokes iLoader and runs a script that tries to unlink two tables. The iLoader script that gets called simply says

DROP LINK [DATABASE].[TABLE].[COLUMN] [DATABASE].[TABLE_2].[COLUMN_2]

The error I get is "Invalid number of parameters in DROP at line 1."

Unfortunately I don't have a document or resource on iLoader so I can't easily look up the correct format.

Ideas?

Mike

Can't find bundle for base name org.pentaho.di.trans.steps.pentahoreporting.messages.

$
0
0
When using Pentaho Reporting Output step and trying to produce PRD report: Can't find bundle for base name org.pentaho.di.trans.steps.pentahoreporting.messages.messages, locale pl_PL

During this Jira ticket https://jira.pentaho.com/browse/PDI-17280 solution is:

"Create a proper fallback message bundle or do not use the JDK standard methods when inside the Kettle plugin."

But what does it mean?
Regards

Chart dont showup

$
0
0
Hello,
Just install v. 8.1 and just for test i add a chart with sample : it never showup ...



:confused:

Just want to notice that V.7 works great on it !

My config server i7 8Gb Windows 7 SSD450

Data source Wiz : key field blank

$
0
0
Hello,

My config :

Pentaho CE = pentaho-server-ce-8.1.0.0-365
Mysql = Vesta CP
Server Pentaho :
Win server 2008 R2
i7
8Gb RAM
SSD124


I test to connect mySQL and key fields remain blank ...



Here all are ok DB



Here what I see log. (PS = I host a Wordpress also on this server mysql so it should be ok ...)



Pentaho = HARD ! .....
Viewing all 16689 articles
Browse latest View live