Senior Data Science Worker

From GM-RKB
Jump to navigation Jump to search

A Senior Data Science Worker is a data science worker who is a senior expert.



References

2020

  • https://www.glassdoor.com/Job/senior-data-scientist-jobs-SRCH_KO0,21.htm
    • QUOTE:
      1. Work closely with our customers to develop a strategy to extract the maximum possible value from the available data
      2. Be involved throughout customer interaction from project scoping to final delivery
      3. Critically assess the type and quality of customer data and work with them to appreciate their toughest problems
      4. Work independently or with our research team to develop a modelling strategy to sit at the core of a data-led simulation-based approach to problem solving
      5. Take ownership of ensuring and demonstrating the reliability of the models that we create
      6. Drive thought leadership within the company and help our clients to understand their data needs and design their data strategies. This involves keeping up-to-date with a diverse body of cutting-edge research, designing and prototyping pioneering new technical approaches and rapidly developing expertise in new subject areas to support new bids and projects.
      7. Strong background and experience in delivery of technical, data-rich projects ideally for external clients.

2018

  • https://www.datasciencecentral.com/profiles/blogs/10-differences-between-junior-and-senior-data-scientist
    • QUOTE:
      1. Automating tasks. Writing code that writes code.
      2. Outsourcing tasks to junior members or to consultants.
      3. Managing people, hiring the right people, managing managers who report to you.
      4. Training colleagues who might not be tech-savvy. Be an adviser for senior managers.
      5. Identifying the right tools and assessing the benefits and minuses of vendor software and platforms, for a specific large-scale project (construction of a huge taxonomy, etc.)
      6. Identifying the right algorithms and statistical techniques for a specific project. Blending these techniques as needed for optimal performance.
      7. Not trusting data; identifying useful external or internal data sources, blending various data sources while cleaning data redundancies and other data issues.
      8. Identifying the best features, perhaps using ratios or transforming, combining raw features to turn them into better predictors. Usually require a good understanding of the business you are in.
      9. Understanding executive talk, and translating executive requests, questions, concerns, or ideas into successful data science implementations.
      10. Measuring the ROI that you bring to your company; being able to convince executives about your added value (or providing sound explanations if ROI is negative or not perceived as positive, and offering a corrective path.)
      11. Interacting successfully with managers / colleagues / executives / clients of all kinds. Mostly a communication issue.
      12. Having a clear understanding of what will create value for your company or client, and be able to deliver that value in a timely fashion, and consistently, despite internal politics and setbacks.
      13. Being able to assess how much time a complex project will require for completion, what the hurdles and rewards will be, and stay on track regarding deliverables and deadlines.
      14. Being able to suggest, create new projects, convince stakeholders, and manage these projects from start to finish.
      15. Being able to jump-start your own company and manage it successfully.
      16. Making recommendations about data science implementations, help with maintenance and make sure a project that started nicely, does not falter over time.
      17. Knowing what you don't know, and manage to outsource or learn new things (and being able to identify / prioritize the knowledge that you need to acquire or outsource.)
      18. Understanding the business you are in, understanding the vision that executives have in mind, even if not clearly stated by your manager.
      19. Testing, testing, testing. Why fake news persist on Facebook despite all the efforts and many millions of dollars spent to fight this plague? Lack of testing, and/or not being able to figure out what fake news look like (lack of business acumen) may be the cause. Scammers change their tactics all the time; your algorithms should identify and take care of new trends, rather than just being able to detect 20 types of fake news, and missing new types that will come after your solution is implemented. Work with business analysts and IP admins to constantly refine your algorithms. Use robust algorithms and robust features. Correctly measure your success rate. Do meaningful cross-validation: if you train your algorithm on 15 cases of fraud (test data set); will it be able to identify 5 other cases of fraud not in the training set, but showing up in the control set? Avoid over-fitting like plague.
      20. Being able to NOT deliver a perfect solution based on a perfect model, requiring three months of work, when an approximate solution can be obtained in one day. After all, data is messy, so perfect models don't really exist. You are just eroding ROI if you want absolute perfection.
      21. Trust your intuition and gut feelings, but only up to some point. Some questions can be answered without writing any piece of code or testing. Sometimes, simple simulations can go a long way. Same with post-mortem analyses (you can call it analytics forensics.)
      22. Spend time on documenting everything, prioritizing, discussing with stakeholders, and planning. At least one hour a day, usually much more.