Actually it's worse. In addition to not staying on one particular
domain, you are also generating urls with some kind of session id.
Sort of never ending.
You have some broken links, and some redirected ones.
Use Xenu Link Sleuth to crawl the site - add both the www and non www
starting url.
http://home.snafu.de/tilman/xenulink.html
You should have a robots.txt file where you disallow various uris or
uri prefixes.
For instance:
User-agent: *
Disallow: /address_book
Disallow: /login
Disallow: /password_forgottten
First fix your navigation to stay either all on www or all without www.
Fix the broken links.
Add the robots.txt file.
Start Gsitecrawler again, delete the project.
Add it again for the particular domain form (with or withotu www).
Ask to read robtos.txt.
Do not ask to import known urls from Google.
Uncheck the option to crawl files types for images, word documents,
pdf, etc.
Go to Filter > Remove Parameters and add a new line for SFHid .
Also all yrou pages have the same title: SFH
Titles should be unique to each page and reflect what the page is
about.
I have tried to run GSC for yrou site - including botth the www and
non www urls in it, banning some ursl in the absence of a robots.txt
file.
I am not able to make headway with GSC.
I wonder if markup errrors are enough to break the crawler:
http://scentimentsfromtheheart.com/
Sometimes somethign stupid like an opening comment like <!-- which
doesn't get closed will mean that everything from that place one is
ignored. Thus no further links are found.
I don't know yet.
Fix what you can from the site and try again.