Hello, bmcompany:
I have just placed a preliminary version on this location:
http://www.xpde.com/HTMLParser.zip
It just to check if I'm going on the right path, you can open any
report and press the Play button to get some results parsed. It stills
don't ask for a list of optimized pages, this will be done once you
press the Play button. Please, tell me if I'm in the right path and if
you like the interface, I will finish it on monday so we can talk
about it and tune it to get a final version on tuesday.
Regards. |
Request for Answer Clarification by
bmcompany-ga
on
07 Sep 2003 03:16 PDT
Hi there.
Absolutely brilliant - works great - and very speedy too.
Ive tested it with the biggest report i can find - 4 MB and 2500
positions, worked perfectly.
Once you have the positions analysed with only the correct pages, then
you are there.
This may be a bit late to add this to the spec, but is there any
chance that you could produce an extra set of figures - positions with
all pages and positions with just the pages in the dir?
If that is a quick job then that would make for a very useful
additional feature.
And for your next trick - (obviously this will be posted as a seperate
question and priced accordingly) can you think of a way to analyse a
directory with say 50 reports - and produce a summary for each report?
This will be tricky when you have to select the pages ? Just something
to think about if you're interested in taking this further.
But it's looking very good indeed.
Thanks again for your time
|
Clarification of Answer by
joseleon-ga
on
08 Sep 2003 00:38 PDT
Hello, bmcompany:
I have some questions, just to finish the next version:
-Can you send me your biggest file, so I can test the performance in
the worst case?
-Can you send me a report and a list of optimized pages that match
that report?
-Do you want to have excluding or including count operations? That is
-If I find a #1 position, right now, this position counts both for:
Number ones
Top 10
Top 20
-Do you want to work this way, or do you want number 1 positions,
just count as number 1 positions and NOT as Top 10 and Top 20?
-Can you send me an HTML template (just a basic design) to put
results inside? If not, don't worry, I will design a basic one
-What would be the default filename to save the results page? i.e. If
the source input file is example1.htm, then, the results can be
results_example1.htm
-Regarding the last requested feature, do you want to add an extra
set?, that is:
First set (optimized pages)
-Total number of positions
-Total number 1 positions
-Total Top 10 positions
-Total Top 20 positions
Second set (all pages)
-Total number of positions
-Total number 1 positions
-Total Top 10 positions
-Total Top 20 positions
Is this what you want?
-Regarding the summary feature you want, no problem, once I finish
the process with a single file, it's just a matter to automate the
process for several files.
Regards.
|
Request for Answer Clarification by
bmcompany-ga
on
08 Sep 2003 02:36 PDT
-Can you send me your biggest file, so I can test the performance in
the worst case?
http://www.sesuk.net/big.zip (link will work in half an hour)
-Can you send me a report and a list of optimized pages that match
that report?
http://www.sesuk.net/pages0001.zip (link will work in half an hour)
-Do you want to have excluding or including count operations? That is
-If I find a #1 position, right now, this position counts both for:
Number ones
Top 10
Top 20
-Do you want to work this way, or do you want number 1 positions,
just count as number 1 positions and NOT as Top 10 and Top 20?
For a #1 position, it needs to be counted as a #1, a top 10 and a top
20.
For a #7 position, it needs to be counted as a top 10 and top 20. Hope
that’s clear.
-Can you send me an HTML template (just a basic design) to put
results inside? If not, don't worry, I will design a basic one
Basic one is fine. This report is only for us to pull positions out
of, not to send to clients.
-What would be the default filename to save the results page? i.e. If
the source input file is example1.htm, then, the results can be
results_example1.htm
Is there any chance that the report could be named based on the first
URL in the report? The report header includes the URL under "Performed
for: www.domain.com" If there are more than one domain, just the first
one is fine.
Additionally could you insert either a random or incremental figure at
the END of the filename?
Hope all that's ok.
-Regarding the last requested feature, do you want to add an extra
set?, that is:
First set (optimized pages)
-Total number of positions
-Total number 1 positions
-Total Top 10 positions
-Total Top 20 positions
Second set (all pages)
-Total number of positions
-Total number 1 positions
-Total Top 10 positions
-Total Top 20 positions
Is this what you want?
Thats exactly correct!
-Regarding the summary feature you want, no problem, once I finish
the process with a single file, it's just a matter to automate the
process for several files.
Excellent, we'll chat about this once this question is closed.
Thank you again for your time and attention to this project. We all
look forward to working further with you in the very near future.
|
Request for Answer Clarification by
bmcompany-ga
on
08 Sep 2003 02:44 PDT
Regarding my previous post. The biggest report i can find is at
www.sesuk.net/big.zip
i dont have the pages for that one, but the other
http://www.sesuk.net/pages0001.zip has a 1.5MB report with the pages.
The report is called report_here.htm.
Thanks again
|
Clarification of Answer by
joseleon-ga
on
08 Sep 2003 02:48 PDT
Hello bmcompany,
I have just uploaded an updated version at the same time you were
answering my clarification, here it is:
http://www.xpde.com/HTMLParser2.zip
I'm going to answer your last clarification:
"http://www.sesuk.net/big.zip (link will work in half an hour)"
Great.
"http://www.sesuk.net/pages0001.zip (link will work in half an hour)"
Great too.
"For a #1 position, it needs to be counted as a #1, a top 10 and a top
20. For a #7 position, it needs to be counted as a top 10 and top 20.
Hope
that’s clear."
Perfectly clear and this is the way it works now.
"Basic one is fine. This report is only for us to pull positions out
of, not to send to clients."
Ok, already included in the last update, please, check it out, in any
case I'm going to externalize the template, so you can change it at
any time if needed, don't worry, it doesn't takes too much time.
"Is there any chance that the report could be named based on the first
URL in the report? The report header includes the URL under "Performed
for: www.domain.com" If there are more than one domain, just the first
one is fine."
I will check it out, in the last update is named results_XXXXX.htm, I
will change it.
"Additionally could you insert either a random or incremental figure
at
the END of the filename?"
A timestamp it's ok for you? Visible or not visible? Readable or a
unix-kind timestamp?
"Thats exactly correct!"
Ok, already included on the last update, please, check it out.
"Excellent, we'll chat about this once this question is closed.
Thank you again for your time and attention to this project. We all
look forward to working further with you in the very near future."
Thanks, you are a very kind customer ;-)
Regards.
|
Request for Answer Clarification by
bmcompany-ga
on
08 Sep 2003 03:19 PDT
a visable timestamp is fine.
Ive tested the latest version and it's 100% perfect! All we wanted and
more.
If you could sort the report naming of the report with the domain that
would be a great bonus, but if its gonna take a while i understand
that you have worked very hard on this already and we already have
more than we asked for.
A million thanks again.
|
Request for Answer Clarification by
bmcompany-ga
on
08 Sep 2003 03:38 PDT
We've just spotted something.
can you investigate this for me?
http://www.sesuk.net/possible_bug.zip
The total results reads LESS than the optimised results. I thought at
first that the results were just the wrong way round. ie,
optimised/total - but it works fine with the other reports.
Let me know what you think
|
Clarification of Answer by
joseleon-ga
on
08 Sep 2003 04:06 PDT
Hello, bmcompany:
The problem was that optimized pages are named like this:
London_Connaught.htm
The_London_Connaught.htm
So when I was looking into an url that pointed to
The_London_Connaught.htm, it was counted twice because it also matched
London_Connaught.htm. It's fixed now, download it from:
http://www.xpde.com/HTMLParser3.zip
I will fix the rest of things this afternoon. Also, I would like you
do the following test:
-Get an small report but which features most of the ways a report
can be
-Then, calculate *by hand* what would be the results
-Then, use the software and see if the results match
If not, send it to me, to check it, this is the only way you can be
sure the software works as it should, maybe I'm forgetting to count
some kind of link.
Regards.
|
Request for Answer Clarification by
bmcompany-ga
on
08 Sep 2003 05:06 PDT
http://www.sesuk.net/checker1.zip
It is slightly out with this one. Some postions arnt being counted. I
think it's because a keyword has 2 positions. EG, listed at #1 AND #7.
have a look and let me know what you think.
|
Clarification of Answer by
joseleon-ga
on
08 Sep 2003 06:40 PDT
Hello, bmcompany:
I have fixed the problems, it was just a matter to add more types of
lines to parse, feel free to download it from here, also included all
the features we were talking about (templates, timestamp, reportname,
... etc)
http://www.xpde.com/HTMLParser4.zip
As soon as you tell me everything it's ok, I will post the source
code.
Regards.
|
Request for Answer Clarification by
bmcompany-ga
on
08 Sep 2003 06:55 PDT
Ok, all is good!
Please post the source code and ill accept the answers.
I think this answer easily deserves 2 stars!
|
Request for Answer Clarification by
bmcompany-ga
on
08 Sep 2003 07:09 PDT
Im only joking about the 2 stars:)
2 small problems to report.
1. the save as doesnt work. The button does nothing.
2. Copying from the results preview doesnt work. The option to copy
appears from the right-click, but the data isnt copied to the
clipboard.
|
Clarification of Answer by
joseleon-ga
on
08 Sep 2003 07:11 PDT
Hello, bmcompany:
You can download the source code from here:
http://www.xpde.com/HTML_Parser.zip
I have used Delphi 7, but it will compile easily with Delphi 5,6. If
you have any problems, just tell me. Also, if you need more comments
on the source code (I think there isn't because the code it's very
easy to read), just tell me also.
Regarding the poor rating you mention, I expect something more than 2
stars!!!!! ;-)
Regarding any other feature you want, you are free to post a question
for any researcher, but if you are interested in me, you can place in
the subject, For joseleon only.
Also, if you find a bug, please, don't hesitate to contact with me.
Regards.
|
Clarification of Answer by
joseleon-ga
on
08 Sep 2003 07:28 PDT
Hello, bmcompany:
"Im only joking about the 2 stars:)"
I hope so ;-))
"1. the save as doesnt work. The button does nothing."
Please, be sure you are using the last version I sent you, that option
was disabled on previous versions. Also, be sure the button graphic
it's enabled, that happens when you load a report into the tool.
"2. Copying from the results preview doesnt work. The option to copy
appears from the right-click, but the data isnt copied to the
clipboard."
That's really strange..., do you have any clipboard management utility
or something similar? I have just tested it and it works ok. Please,
tell me the size (in bytes) of the HTMLParser.exe file you are using,
please.
Regards.
|