Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

Input from Excel 2013

$
0
0
Hello Pentaho forum!

I have recently begun a proof of concept using Pentaho Data Integration - CE 5.3.0.0.
I'm trying to load various format files including csv, xls and xlsx into an Oracle database, and am having issues reading from Excel 2013 sheets.
There's no explicit 'Spreadsheet type (engine)' listed for Excel 2013, does that mean I need to install an add on? Or is this just not yet available?
If I save the Excel 2013 sheet as xls I can read it fine, however I would like not to have to use non-native file formats if possible.

Thanks in advance,
James.

Report Creation with Java very slow

$
0
0
Hi there,


while trying to create reports with Pentaho and Java I discovered, that the creation process is very slow and memory-expensive. I created a prpt-file which I use to display the data in. Also, I have a factory class which returns the TableModel (DefaultTableModel) in which the data is stored.


While with the Pentaho Report Designer everything seems to work fine, with my Java program it's very slow and uses a lot of memory with more data.

I'm using this method to create the PDF file:
Code:

PdfReportUtil.createPDF(reportObject, fileObject);
Sure, I have a lot of data and the PDF is very large (about 4000 pages) but 50 minutes seem a lot of time to me. Or isn't it? Is it even possible with Pentaho to handle such amount of data?

I also have problems with other formats of the report. With RTF or XLSX I'm getting OutOfMemory exceptions.

I'm using Pentaho Designer 5.3.0.0-213 and the corresponding Java libs.

Thanks for reading
s3b

Row to column

$
0
0
Hi, I have one sample input file- file.txt(attached) and I want sample.txt(attached) as output. Is this possible using row denormaliser or any other step??

Thanks in advance for your support.

Regards,
Poulomi
Attached Files

how to get Analysis report option in pentaho biserver community edition

$
0
0
hi

i installed pentaho bi server ce 5.0 on windows, but i am unable to get analysis report option on console

pls forward any one knows.

Thanks

srinivas

Unexplainable increasing inactivity every ExecutionPhaseEvent

$
0
0
Hi all,

I've been looking into the mondrian log file because of a query execution that doesn't seems to finish.
I between every ExecutionPhaseEvent and SqlStatementEndEvent an increasing time span where I don't receive any logging.

Code:

2015-04-01 15:39:09,774 DEBUG [mondrian.server.monitor] SqlStatementEndEvent(78)
2015-04-01 15:39:41,527 DEBUG [mondrian.server.monitor] ExecutionPhaseEvent(790, 58)

Here we see it takes 30 seconds to start another ExecutionPhase, and this number increases every time.

Could somebody explain this behavior?

CSV Loader with British Summer Time

$
0
0
Hello!

I have my forecast set up and running in Java, getting the results I needed. However the clocks went forward in the UK on the 29th, so the times between 1am and 2am on that day don't technically exist... so I'm getting an 'unparseable date' error on those timestamps...

Any idea how to fix this? I thought a skip list could work but that's defined after the data is loaded so after the error is thrown!

Thanks in advance!

Pentaho Connectivity to Hortonworks Sandbox HDP 2.2 Hive tables

$
0
0
Could you please let me know how to connect Pentaho Analyzer or Pentaho Reporitng Designer to Hortonworks HDP2.2 hive tables.

Using parameters in database connections

$
0
0
Hi

We're using parameters for our database hosts in Kettle. This works great, and we can pass the parameters into kitchen on the command line.

However, we're hitting an error in Spoon when we try to test things locally. If we do a "Read Fields", for example, we get an error that the host ${MYSQL_HOST} is not found. So I don't think that "Read Fields" is able to properly substitute the parameter here.

If we run the transform, it works fine and substitutes the parameter. It just seems to be doing a "Read Fields" where things fall down...

Thanks,

Tom

Need help with Id3 Classifier

$
0
0
Hello I'm trying to reconfigure the Id3 tree classifier to do the following three things;

1)Consider a probabilistic version of the ID3 algorithm that proceeds as follows. At each step, theinformation gain of each potential pivot is computed. I then select the pivot probabilisticallybased on a roulette wheel equation

2)Now consider the Ensemble probabilistic ID3 algorithm (called EpID3) algorithm which proceedsby running the previous change M times, where M is an external parameter. This willproduce an ensemble of M (potentially different) trees.

3)Each new unlabelled instance x would be applied to each of these M trees where each tree would give it a class label. Then the class label that occurred most frequently would be the output of the algorithm in the form of a probability score where the probability score can be interpreted as the algorithm saying "theunlabeled instance belongs to the class c with probability y"(for example)

I've implemented part 1 easily enough as just a simple edit of the Id3.java file, but I'm having difficulties find out where in the source to alter the number of tree created and how they operate. I fairly certain it would be in the ClassifierPanel.java but the code is a little too complicated for me to understand exactly how the classifier object is working and if I could perhaps make an array of them. The array would contain each of the M trees created and then I think it would be simple enough to get the rest of the algorithm from there.

If anyone here is knowledgeable on altering the source code of WEKA and could give me a hand I would really appreciate it.

CacheControl.flushSchema & deadlock

$
0
0
Hi.
Sometime, we have the problem with clear cache of OlapConnection and executing mdx query in parallels threads.
All olapStatements waiting the next thread:

"http-bio-64.187.229.242-443-exec-166" - Thread t@1752

java.lang.Thread.State: WAITING
at sun.misc.Unsafe.park(Native Method)
- parking to wait for <1aced1d6> (a java.util.concurrent.CountDownLatch$Sync)
at java.util.concurrent.locks.LockSupport.park(Unknown Source)
at java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(Unknown Source)
at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source)
at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source)
at java.util.concurrent.CountDownLatch.await(Unknown Source)
at mondrian.util.SlotFuture.get(SlotFuture.java:87)
at mondrian.olap.Util.safeGet(Util.java:2095)
at mondrian.util.BlockingHashMap.get(BlockingHashMap.java:63)
at mondrian.rolap.agg.SegmentCacheManager$Actor.execute(SegmentCacheManager.java:1059)
at mondrian.rolap.agg.SegmentCacheManager.execute(SegmentCacheManager.java:355)
at mondrian.rolap.agg.AggregationManager$1.flushNonUnion(AggregationManager.java:133)
at mondrian.rolap.CacheControlImpl.flushInternal(CacheControlImpl.java:205)
at mondrian.rolap.CacheControlImpl.access$000(CacheControlImpl.java:37)
at mondrian.rolap.CacheControlImpl$1.execute(CacheControlImpl.java:180)
at mondrian.rolap.CacheControlImpl$1.execute(CacheControlImpl.java:178)
at mondrian.server.Locus.execute(Locus.java:92)
at mondrian.server.Locus.execute(Locus.java:77)
at mondrian.rolap.CacheControlImpl.flush(CacheControlImpl.java:175)
at mondrian.rolap.agg.AggregationManager$1.flush(AggregationManager.java:156)
at mondrian.rolap.RolapSchema.flushSegments(RolapSchema.java:324)
at mondrian.rolap.RolapSchema.finalCleanUp(RolapSchema.java:346)
at mondrian.rolap.RolapSchemaPool.remove(RolapSchemaPool.java:390)
at mondrian.rolap.RolapSchemaPool.remove(RolapSchemaPool.java:380)
- locked <48af73ea> (a mondrian.rolap.RolapSchemaPool)
at mondrian.rolap.CacheControlImpl.flushSchema(CacheControlImpl.java:301)

This is deadlock :-(.

The way in which we must to debugging mondrian code?

We are using mondrian 3.9.

IBM DashDB

$
0
0
Hello All,

Does anybody know if Pentaho Analytics or Data Integration support connection to IBM DashDB?

Thanks

IBM DashDB connection

$
0
0
Hello All,

Does anybody know if Pentaho Analytics or Data Integration support connection to IBM DashDB?

Thanks

Table input converts postgres numeric(7,2) to kettle number with precision 7,2

$
0
0
We have a postgres table with numeric(7,2) currency fields:

Code:

                                                Table "foo"
            Column            |            Type            |                                      Modifiers                                     
------------------------------+-----------------------------+-------------------------------------------------------------------------------------
 cost                        | numeric(7,2)                |
 revenue                      | numeric(7,2)                |

We use a Table Input step to load rows from this table. Looking at "Show Output Fields" from that step, I see these fields are converted to a kettle Number -- which is a lossy Double -- instead of to an precise BigNumber.

I've tried hinting Kettle to do the right thing using an explicit cast within the Table Input sql but that doesn't help.

Code:

SELECT
  cost::numeric(7,2),
  revenue::numeric(7,2)
from
  foo

..but that doesn't help.

Does anyone have a suggestion on how to deal with this?

Dimension and fact tables

$
0
0
Hi everyone, I really need ur help.

I'm a beginner in BI, I designed a DW and I want to implement it using KETTLE PDI and ORACLE database.
I created tables in my database as DIMENSIONS and another one as FACT TABLE.
To load my dimensions i'm using the "extract from table" ( from the source table in my database) and "insert into table" ( the dimension table that i created in my database containing the ID from the source table and another technical ID auto increment).

I've some questions :
1) Am I on the right way, or there is better ways to create dimension and fact tables?
2) Can u please explain what's the difference between ''insert into table'' and "slowly changing dimension", do I need to use slowly "changing dimension" or a simple insert into my dimension table is enough.
3) How can I load my fact table, with all the foreign keys from dimensions and also calculate all the measures in KETTLE.

Thank you for your help.

Doubt Regarding pentaho data integration

$
0
0
Hi,

I have a doubt regarding Kettle.
Suppose i have simple transformation having a table input step.
If i open the transformation in my PDI and then open that table input step and then i do edit connection details.
I modify the connection details and do test connection without doing ok step to save that change.
Now my doubt is when i am testing the new connection details without doing the OK button,it is going to test the old connection also which is changed
but not permanently removed as i haven't pressed OK.

Is it like the old connection is still there in buffer and that also tested in the background when i am testing the modified one.

Please help, its very important.

Thanks in advance

tooltip problem

$
0
0
I've a report with a tool tip on a text-field. Everythings works fine on my PC when I use the report designer.
The tool tip doesn't appear via BI Server when I publish the report.
Both report designer and BI server are 5.3.
Any idea?
Thanks
Matteo

tooltip problem - PRD 5.3 & BI server 5.3

$
0
0
I've a report with a tool tip on a text-field. Everythings works fine on my PC when I use the report designer.
The tool tip doesn't appear via BI Server when I publish the report.
Both report designer and BI server are 5.3.
Any idea?
Thanks
Matteo

elipse disapear in HTML (single page) report - PRD 5.3 & BI Server 5.3

$
0
0
I've an elipse. The fill color property is linked to some value using a formula (like a traffic light).
=IF(OR([ERRORE_MANCA]=1;[ERRORE_ADD]=1);"red";IF([WARNING]=1;"yellow";"green")).
When I display the report in HTML (paginated) I saw the elipse, but choosing the single page format they disappear
Thanks
Matteo

Format Label Text in Bubble chart in Pentaho Report Designer

$
0
0
Hi,

I want to show legends as the bubble labels in the bubble chart where legends and y,z axis refers to the different columns of the data source (legends refer to A column(string type) and y,z axis refer to B column(number type).

Is their any method to do this?

BI Server 5.3 and Active Directory

$
0
0
Hi, I use BI Server 5.3 and trying to set up authentication through Active Directory, but it does not work. I use instructions

Catalina log:

tail -f tomcat/logs/catalina.out
Apr 02, 2015 3:01:20 PM org.apache.jk.server.JkMain start
INFO: Jk running ID=0 time=0/15 config=null
Apr 02, 2015 3:01:20 PM org.apache.catalina.startup.Catalina start
INFO: Server startup in 4965 ms
Apr 02, 2015 3:03:16 PM org.apache.coyote.http11.Http11Protocol pause
INFO: Pausing Coyote HTTP/1.1 on http-8080
Apr 02, 2015 3:03:17 PM org.apache.catalina.core.StandardService stop
INFO: Stopping service Catalina
Apr 02, 2015 3:03:17 PM org.apache.coyote.http11.Http11Protocol destroy
INFO: Stopping Coyote HTTP/1.1 on http-8080
Apr 02, 2015 3:03:32 PM org.apache.catalina.core.AprLifecycleListener init
INFO: The APR based Apache Tomcat Native library which allows optimal performance in production environments was not found on the java.library.path: /usr/java/packages/lib/amd64:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib
Apr 02, 2015 3:03:32 PM org.apache.coyote.http11.Http11Protocol init
INFO: Initializing Coyote HTTP/1.1 on http-8080
Apr 02, 2015 3:03:32 PM org.apache.catalina.startup.Catalina load
INFO: Initialization processed in 545 ms
Apr 02, 2015 3:03:32 PM org.apache.catalina.core.StandardService start
INFO: Starting service Catalina
Apr 02, 2015 3:03:32 PM org.apache.catalina.core.StandardEngine start
INFO: Starting Servlet Engine: Apache Tomcat/6.0.41
Apr 02, 2015 3:03:32 PM org.apache.catalina.startup.HostConfig deployDescriptor
INFO: Deploying configuration descriptor pentaho.xml
log4j:WARN Continuable parsing error 187 and column 23
log4j:WARN The content of element type "log4j:configuration" must match "(renderer*,throwableRenderer?,appender*,plugin*,(category|logger)*,root?,(categoryFactory|loggerFactory)?)".
[Server@46ae10a6]: [Thread[main,5,main]]: checkRunning(false) entered
[Server@46ae10a6]: [Thread[main,5,main]]: checkRunning(false) exited
[Server@46ae10a6]: Initiating startup sequence...
[Server@46ae10a6]: Server socket opened successfully in 1 ms.
[Server@46ae10a6]: Database [index=0, id=0, db=file:../../data/hsqldb/sampledata, alias=sampledata] opened sucessfully in 1381 ms.
[Server@46ae10a6]: Database [index=1, id=1, db=file:../../data/hsqldb/hibernate, alias=hibernate] opened sucessfully in 19 ms.
[Server@46ae10a6]: Database [index=2, id=2, db=file:../../data/hsqldb/quartz, alias=quartz] opened sucessfully in 26 ms.
[Server@46ae10a6]: Startup sequence completed in 1429 ms.
[Server@46ae10a6]: 2015-04-02 15:03:35.971 HSQLDB server 1.8.0 is online
[Server@46ae10a6]: To close normally, connect and execute SHUTDOWN SQL
[Server@46ae10a6]: From command line, use [Ctrl]+[C] to abort abruptly
Pentaho BI Platform server is ready. (Pentaho Open Source BA Server 5.3.0.0-213) Fully Qualified Server Url = http://localhost:8080/pentaho/, Solution Path = /opt/biserver.5.3.postgresql/pentaho-solutions
15:04:02,581 ERROR [CteDefaultProviderManager] Provider ID is blacklisted: sparkl. Discarding it..
Apr 02, 2015 3:04:03 PM org.apache.catalina.startup.HostConfig deployDirectory
INFO: Deploying web application directory sw-style
Apr 02, 2015 3:04:03 PM org.apache.catalina.startup.HostConfig deployDirectory
INFO: Deploying web application directory ROOT
Apr 02, 2015 3:04:03 PM org.apache.catalina.startup.HostConfig deployDirectory
INFO: Deploying web application directory pentaho-style
Apr 02, 2015 3:04:03 PM org.apache.coyote.http11.Http11Protocol start
INFO: Starting Coyote HTTP/1.1 on http-8080
Apr 02, 2015 3:04:03 PM org.apache.jk.common.ChannelSocket init
INFO: JK: ajp13 listening on /0.0.0.0:8009
Apr 02, 2015 3:04:03 PM org.apache.jk.server.JkMain start
INFO: Jk running ID=0 time=0/29 config=null
Apr 02, 2015 3:04:03 PM org.apache.catalina.startup.Catalina start
INFO: Server startup in 30948 ms
Attempting to load ESAPI.properties via file I/O.
Attempting to load ESAPI.properties as resource file via file I/O.
Not found in 'org.owasp.esapi.resources' directory or file not readable: /opt/biserver.5.3.postgresql/tomcat/bin/ESAPI.properties
Not found in SystemResource Directory/resourceDirectory: .esapi/ESAPI.properties
Not found in 'user.home' (/home/pentaho) directory: /home/pentaho/esapi/ESAPI.properties
Loading ESAPI.properties via file I/O failed. Exception was: java.io.FileNotFoundException
Attempting to load ESAPI.properties via the classpath.
SUCCESSFULLY LOADED ESAPI.properties via the CLASSPATH from '/ (root)' using current thread context class loader!
SecurityConfiguration for Validator.ConfigurationFile not found in ESAPI.properties. Using default: validation.properties
Attempting to load validation.properties via file I/O.
Attempting to load validation.properties as resource file via file I/O.
Not found in 'org.owasp.esapi.resources' directory or file not readable: /opt/biserver.5.3.postgresql/tomcat/bin/validation.properties
Not found in SystemResource Directory/resourceDirectory: .esapi/validation.properties
Not found in 'user.home' (/home/pentaho) directory: /home/pentaho/esapi/validation.properties
Loading validation.properties via file I/O failed.
Attempting to load validation.properties via the classpath.
validation.properties could not be loaded by any means. fail. Exception was: java.lang.IllegalArgumentException: Failed to load ESAPI.properties as a classloader resource.

help me please.
Viewing all 16689 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>